var/home/core/zuul-output/0000755000175000017500000000000015136702615014533 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015136707113015475 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000252162515136706737020302 0ustar corecoreߍ{ikubelet.log_o[;r)Br'o b-n(!9t%Cs7}g/غIs,r.k9GfD ~3Eڤ펯_ˎ6Ϸ7+%f?長ox[o8W56!Kޒ/h3_.gSeq5v(×_~^ǿq]n>߮}+ԏbś E^"Y^-Vۋz7wH׋0g"ŒGǯguz|ny;#)a "b BLc?^^4[ftlR%KF^j 8DΆgS^Kz۞_W#|`zIlp_@oEy5 fs&2x*g+W4m ɭiE߳Kfn!ׯ*A(-aD~JwFPO7M$n6iXύO^%26lDt#3{f!f6;WR.!$5 J:1*S%V!F([EbD]娍ԹiE03`Cfw&:ɴ@=yN{f}\{+>2^G) u.`l(Sm&F4a0>eBmFR5]!PI6f٘"y/(":[#;`1}+7 s'ϨF&%8'# $9b"r>B)GF%\bi/ Ff/Bp 4YH~BŊ6EZ|^߸3%L[EC 7gg/碓@e=Vn)h\\lwCzDiQJxTsL] ,=M`nͷ~Vܯ5n|X&pNz7l9HGAr Mme)M,O!Xa~YB ɻ!@J$ty#&i 5ܘ=ЂK]IIɻ]rwbXh)g''H_`!GKF5/O]Zڢ>:O񨡺ePӋ&56zGnL!?lJJYq=Wo/"IyQ4\:y|6h6dQX0>HTG5QOuxMe 1׶/5άRIo>a~W;D=;y|AAY'"葋_d$Ə{(he NSfX1982TH#D֪v3l"<, { Tms'oI&'Adp]{1DL^5"Ϧޙ`F}W5XDV7V5EE9esYYfiMOV i/ f>3VQ 7,oTW⇊AqO:rƭĘ DuZ^ To3dEN/} fI+?|Uz5SUZa{P,97óI,Q{eNFV+(hʺb ״ʻʞX6ýcsT z`q 0C?41- _n^ylSO2|'W'BOTLl-9Ja [$3BV2DC4l!TO C*Mrii1f5 JA *#jv߿Imy%u LOL8c3ilLJ!Ip,2(( *%KGj   %*e5-wFp"a~fzqu6tY,d,`!qIv꜒"T[1!I!NwL}\|}.b3oXR\(L _nJB/_xY.# ſԸv}9U}'/o uSH<:˷tGLS0l/LKcQ.os2% t)Eh~2p cL1%'4-1a_`[Zz㧦|k˭c ĚOρ_} Ewt3th?tvͪ{~;J0= |JUԍ;Iw}/9nh7l%>'ct Հ}a>-:(QxPyA Z UcÖgڌ:8cΗ|U1,-N9 dI [@3YN%:ò6PT:”QVay 77ĐrX(K&Y5+$wL#ɽ 4d-bbdAJ?w:P>n^2] e}gjFX@&avF묇cTy^}m .Ŏ7Uֻ󂊹P-\!3^.Y9[XԦo Έ')Ji.VՕH4~)(kKC&3[n )ܗKj/jUSsȕD $([LH%xa1yrOpY]Q4`Iz_*2coT'ƟlQ.Ff!bpRw@\6"yr+i37Z_j*YLfnYJ~Z~okJX ?A?gU3U;,ד1t7lJ#wՆ;I|p"+I4ˬZcն a.1wXhxDI:;.^m9W_c.4z+ϟMn?!ԫ5H&=JkܓhkB\LQ"<LxeLo4l_m24^3.{oɼʪ~75/nQ?s d|pxu\uw?=QR -Mݞίk@Pc n1æ*m$=4Dbs+J \EƄզ}@۶(ߐ/ۼ𹫘qݎt7Ym݃|M$ 6.x5 TMXbXj-P\jА޴y$j`ROA"EkuS#q * CƂ lu" yo6"3껝I~flQ~NCBX`]ڦÞhkXO _-Qy2$?T3ͤEZ긊mۘ$XD.bͮW`AީClСw5/lbl[N*t*@56."D/< {Dۥ sLxZn$N(lYiV =?_e^0)?]{ @| 6+#gPX>Bk2_@L `CZ?z3~ }[ tŪ)۲-9ֆP}b&x Uhm._O 4m6^^osVЦ+*@5Fˢg'!>$]0 5_glg}릅h:@61Xv` 5DFnx ˭jCtu,R|ۯG8`&ו:ݓ3<:~iXN9`2ŦzhѤ^ MW`c?&d.'[\]}7A[?~R6*.9t,綨 3 6DFe^u; +֡X< paan}7ftJ^%0\?mg5k][ip4@]p6Uu|܀|Kx6خQU2KTǺ.ȕPQVzWuk{n#NWj8+\[ ?yiI~fs[:.۽ '5nWppH? 8>X+m7_Z`V j[ s3nϏT=1:T <= pDCm3-b _F(/f<8sl, 0۬Z"X.~b٦G3TE.֣eմi<~ik[m9뀥!cNIl8y$~\T B "2j*ҕ;ێIs ɛqQQKY`\ +\0(FęRQ hN œ@n|Vo|6 8~J[,o%l%!%tyNO}}=ʬ-'vlQ]m"ifӠ1˟ud9)˔~BѤ]һS8]uBi( Ql{]UcLxٻa,2r(#'CDd2݄kTxn@v7^58þ Ţ&VY+yn~F8I !6WB3C%X)ybLFB%X2U6vw8uUF+X|YukXxVO(+gIQp؎Z{TcR@MSRδ~+1æ|mq՗5$B᲋eY(|*磎\Dži`dZe j'V!Mu@ KV{XץF .Jg< ƜINs:b zĄu3=Az4 u5'og^s7`Rzu-anOIq;6z( rx߅ euPvIɦ7聀t>G;_H;2ʗ6 h6QװxmR JQUbTP2j˔Ni)C)HKE"$ӝ!@2<Bq 2oh80,kNA7,?ע|tC3.㤣TiHEIǢƅaeGF$ u2`d)/-st{E1kٌS*#¦۵_Vu3ЩpRIDr/TxF8g4sѓ{%w .ʕ+84ztT:eEK[[;0(1Q@ET0>@wY)aL5ׄӫ A^%f+[`sb˟(]m`F3 W((!5F-9]dDqL&RΖd}})7 k11 K ;%v'_3 dG8d t#MTU']h7^)O>?~?_ȿM4ə#a&Xi`O}6a-xm`8@;of,![0-7 4f kUy:M֖Esa./zʕy[/ݩqz2¼&'QxJE{cZ7C:?pM z*"#窾+ HsOt۩%͟A498SwWv|jNQ=-[ӓI]mSDzCY`̵"_zfz^DI9ןYIpdǧPoUbCEYmoOwGeXYlSJamK(Vwշy»QU5ٲkm@ B̠Yߒ`rH&C|@$XXg1/ 4sCd0!ֵ9lX #QYd.x3'[sJHt**feYŇ\[S={3CfN:XCJFYE10]@  2OGd.h, :<%uab`~q}sY{x<"~-KU3.cn3CYR_^tR*ULF"k3,K- 5u|1,V% "HDxEV#XFҼ7 $sz& !"uYhy,%oʀs"/b]֗jos==T2gႎ&l<}_5 <M579"cY{RJ'ÚbB>*`yCBA6Z "k+ٞE˾do`:nR</r`kC%X axQfd>$аOW.nVcx}T+ɘtnrǬV`NZZ rPw㈹ <9|@W`| ÚBd wBe 10Vegɼ3Vׯ\^wKg̑F3%6!g|0W*xi3EXǛI{҇7}.P|pTzzzh9~KYvxH]09;}?:k_vR `p=ڣ߆ A$@p=VwcO;"s(e QFp8ܑMw*;CMt]﮶r{00ya}N{/^c!)X>[S:#J5Pl4Iޭ 3 YN 4^]#cc2:P^*%ɻYHr=pXTAĤQ(QG,]d9 ̹jjB&d0fE!8)' ;8و9Ș&NhCq2 PG~g_D40;ҼG}d]4ahN/9dutt`D^.rP g6) ]>O[XZkYFEU6#몗FQJ)eYGŇԅm0}VSWUu_Yۛ&UQnW}y!d}{,iamatP>B$")om2Ja&ڎDygŖ.рܩB) X8"#"ېNCe ̓c@;YHYb]yһhRї';}ht׃7}˳ӷoϚ[%ݥƃJ' yJ~ภZ?{՞)u#ZOĠJǻomd!cwG2"*jOEeSq )!kUi͉xA%Y"k n^ hb$lsYMPx*/vf?*/k5}p> @YNJ{T-I^lbVŬ-{\06iܶ.}4 %ڑ(Ra>CʤxO7zE4AL.NӇd@7#]#e,&|FL yw4^%MSw?rnN'$ks" q#Ku&?tNGO_^\.\OiPOΉ:KџxTyRMZe g͎Tx]荞{!usW}2_cld)!sH>=z! Ys5F7-LC(l ϠKCcX29FԝE^eܙ_~;:}0_~(t}J6EuyC'֪$exW#E4_if>SߴЅfm_m;j OsV`!/Y#gѡ$oy)kKdD}}ֵ>/[aY$Qђ2kx%atiˍ`y,$|yl .2~nz['afl8!4c*2H?K;P#;99֪ġ.շv VF5H̪u|doV kH HYp5<gD+(@N[tZ 9AI35 ZMH>Y 2FQ5DG2k3Zb g@b^U6~_I[y@ڮr>(:D;'Q40EKK%(U܃9ʬcu,7N:N^!X_*Aa:y+27fx-y"Z/2{":rUa a}6 3Nmhkc]r~?N2O8_ n[v:d 9NKG}3-辽auaS7mվnU,NY͆Cm GE 7HbgYj筂kݼOǧ~f-^HEKg_n_u;޴'{+8몲]O/[U>Yjɰ&-H;:ӫߏ. l w]l:0w-קѤʀwpf`"Ve_x4`T*t"*LH&=-r9Gֈ Nd6Ⱥ'D@@ :U]:7tld{ g"  66 <[Sta}"Hjk+ZG8SnË.w`޷@N$@W,2p/a]EuC\baݪtiT>~ml-rȚ)L")WeY}-6ȱ 9/+T4H{#@;S_*Y:, o[_o`3goSΒwQIX9܊[+=ku!Au +5?dtK}&XxM!xú~޽>R0a`fق҉5%>\X %LSa2Ϟ}lJl>/>+s2yPH& P1dR_']i6#wwOk{. Ӵ͚M~Ӿ%cم`wT<4EþNqL2H/2JF"EcNM4M`ڔ I}v/jc +8P * hkMVV{|%_}ػˋ]^s]dw4Yt[X++s{?wyvV=M|o;4bj; RTJ{d;W$^ӇSW}4,m! zK4rdeʚƀg]XaE/Y]Fj=/ V։Il4UhDƐ5=n?>%(~dUE*[!`Mfhd)TcMp`B ݰj(/@cC|5Yo|B)vL,T@1A&[&V rʧ%IOO#[trqQxNU7IȆGlsp?\:k>:63}5͚J>_(vb@3׿CRxbHJHeNni5(~p>O@[I+'["3.FIdKO Y)\2&<~cVѢ)Y[wQ ۊ*@:\1,j!v<}Cj#kZ"DUIowU>I Ie)ݒ&CD%TVF0!mT1+4ĔJC.E'JiJ_mSz)p?uݹGu(9+1>јf^H!ZuL}>cd}5djp(ňًf݃kf=dicd.5&l"o QRT) SyZ?Y-Ƴ]7|Ss/]t47 >W֫~tE\4zt~V$׃%] Z׽KN羙OC>@ jexQJPѥ elBʥ%74dQ ?%}G}ݸwfz=waL6zE\/A Z^rpo^'_Oj<BH|_2b~fy]S%)V}m|X!lk҃pnU5:ZoXZlwoibclůd >04)6fbmPTi-"] /QO*(pco:{*CL…5v,V:{[g./IBfJ9u+Q·B"Q F>wTLAtUG>)d 8scM6MY 6ses~}3ɱ7¹ R$QӅ#tӚe~=W|X-9eb| X1lmU o玳,Ru!`:m:S5PAxtL)Φ *<:1VArӍ?LT%l)= Κz[Ư +RVDžub`WlUai,|Fߟ|. d3EjFi .}&bXԈedKX?uV:fhiqgTĔO5 ĩRVD7ֶ].coa@>RX=4OZS׹Un %xccDa.E h :R]cmuNVC'xZ㣪dC$1aH_d$1'/8[Z4 hEݸDt#{"dq1v>jkT9F-)L)ŘF"uVK j0=7ŰrԋwEDYp{[q`Ȝ/{ 9a>E)XOS%EGƶ*:8DO,FC#zNzySsTtFF> z+KVbZ˙#G6 `8U֕T+g_դ}Y({ -.DaPqobb,n 2w K1,jX4W->L!Uc3Fdk4ե}A>d2UؖvlX6|\!O*/m}#r*E":vSz]k/I,[,|vg>=u^\J+Ȫcv/#4RX :锉voqcU!} /5O߁ݛǪ0q,7F%IP Rڸ7p~ -cl'nRw⠫R"X3 P[ɒkX9U1>V.)W%'X +Uv`=i:ԝ!,Z#Vb rရ\sNC/T"* Q:!9WNSUXM{? 1:)6k)|DN`cv@vDN. aU9O 3u5GY \M뿆Xk3۾ `7LzP9&`|8|Ldl?J21٤MRuy}z{~n\CXߍ|}&zo쟮ͳ,rDY{yM1nu ?.^T卮7sy}X~澼;uwuSkU]_~O-=ί:n'ܸ՟ +[)/:n埢&/8n/!qۅl@Ud_h*m;' `0m!?Z:ZoBeH1Y8J-ޖcS'5Y_䁖%5ғdcY'HA[8kL8E)J\8Ɂ,GMlʂ+lh)1>Z.|>VL h5^7eM>y̆@jCxZ\c| 6?EV"sd!N@БU ^p+Jpᄟ,C'[|BUxY$߾lvi `i2k#@+j۟gWb 1؁c~L :4^ )lp 2\. }B \5>ToEχi)e0\lCMlɱĘR7HebLϧB|+QcX6.v@H98Vctq6H=m*2*q>C!9U;q ^1puOSvBHHmQ (6c94Ce|u5ӄ|  ]وb Xiצ) &5TN]SxL\{n9:3/Y|dH]m-|kۗ~sG4bzgek#+L;:)q!l0/ѝDҥ*Ha4-s>Vz[~) xs%P ع owmWuIC.-} KJI3;j !ҝd`>&5mKMf#pBT+[n*cem:y0W|c0Vvގj$Sp^M+f~D`{be@0o2cވu~-B竟} |6K-uT0\8"M:" /P:3`l' .Z cEpN9K19`ҽFpU۞]tLΦCsy#rFscCbX%E+o*ƾtF*`NΛv7oϯ +Ōa`HDN74Т C>F|$A:XB8dJWqLhnٓԻfl8fp.CDsc3k.2WM:U~{N_>!Lc/rK-vv%~ =WBX"XA:#u-9`x 92$4_!9WvT` d0ϕ_G\ Bؒ G.}ΕU&4D&m.Z9cH,HCԢ{b G-JxD+@_$c%* _jR|\:t >jz(:W֪F\ǔz;m|Ȫ~_A7¸W`uYctmc .I1:Vk7m"cKӝNNǽJOG\VrFU5q:ǫf!NRT1.D(7.8;ІWTSÜ:OUS5^ga7NWQ`1LRx+$_D +'\ U>]BfL6鞞,+Bg#[3`pO^>e7Ds~|h!.5b-EDz3M8V|&jZzˁXnlwì:8Vc1S1:Wh#O?Nm' E [%W%nlꠧ1n<Y^llmb rY״͖jiTLC\?c%0s1:W_j´KxA|Hk6Jb̯["+BLu>Ɩ=xm][4AkE`AW/B$52F$Rv%`[ ضزEM_UA| m.' L,CEaSB]<^w%?-ڏ q`U"oRs?]\MkǴ`+Cc07Ks`hfCk0Q[csK?\g,\ rֲ[.x]io$BѽnTEf3J!dqDzn~+CX|,l_B'9Dcuu|~z+wGqȅlW11/$f*0@б zrt"?ߛ+qry??ڕx2;W`Vj5l0Vm_JL-yz2,Zlu]nc˝߂~[jRuo[|["w;?2Y#tV[GT F*OO '+EoWv4jwB}Ԏѹ .MVfz0ÏABF3O>7-|йJs V[~ۗ#r#r[Fw/Jl?J21&5I媒Ǜou{u{,wߍK벼\MQn%̈v޴kӔ2͆!`co|Lu_~^ǴT?cr_$d`3 P}U=̀*E Zj/u/V ?ç%ra>Ox!8+L򹷬4/.m0E(G-^倕#{ ӡcj{ƕ_1~lz/nOp6ws "D֍$zqw(Ɏ&6n.-p8SٍM?4M8XU"U9Yx*ölur$F $Uw eR%2zDƪ0hF,ƪ$rwAŊ0~ƪIe;,ʤwwM9vX)"ZoHaQ2lʪޞ7w$TGSmX&Bf~jd"O`R<**2nHvf|[cU\HcUU3gF:ɠy: d7 U ƧBOy)钌GT 2-0-zBxyOr.n~اEr3N%q**dp^ %eXKNhŜÈD/>%2Z{r[^W.%v,<~獐?eqe9DرE^WS. <'fܵ\/v"uZ6 w(go 5K؃b(dY| bY*|9N*z5mPQAțl}w{q#!>ɹ)&F]fg߉(rGq2X1[چit/QFM!CrUޏl:&u.MzIKӽ4K˙Z&m׮S4f4*Q J]?b@XI4pPy|dԥzS ґں{}Gz8~dS2x)DSkT-?8V%$FFca /E&ώ*#Q-_#u+O,YSXj̍i[^uURS0Ԉ<T"$YT%^^&*bZp1 K퀡,PE)RE#c<ž[IDh[$ 6eَˉM'sl /G$Q3?2di&!OR9OFr#MLA*St&y 2KV`Ve?$KqK%_/ށtD̀\&j|bc3'ò_|k;w֥3.uM].05nliIR%ap KQ?n,:ϝ9Byka\Qqh5E|yU$,V0MBi`xl\`~6\rx@o/ㇷ_Mw(Dtrqڭ V dSFc :2`KT\ \ʹ(?"o`$X,%oQ|l- x "F{璯k*tO?~ Uy[JW(F-<{*?mԹJUT>tz^9)놥@D 9S6*E9O [_ﳋ%x+E2 ZRKO#k2)[Kj8g(⧾aRB%XZ/N"":%3Dnu{/R4SgI Q("=/Vk3@e\?"1\6EY^PX7M-3M`S1m㗲y0__?yzbs5 GEmupbT5|#Dmm2>áqjoI :muh[:s'3xJ<>^HM흈0 G8R[Z'U-Y 8GiPi:gɆkF vU)),5Qk^;w5W1|8Agupԟ}Ya=ˤ +3 E.c7<`?npFu])?/9,*Ցfgu8ho@]%Íe*X#rf8"P3xfE5*C+ƈmi=Zۘ(ùفpꮹ\ApdҧR5dlֳo{d>%pm "WI­ Vl\CCvE0S4C}.W?k=$|`'!|X0r,j*OOueQAK\MY X͝})w%BJ l1kIO߽t.x@KڕrIy kZ"Y]aNYav*ܔ¨KWI^;:3廻Pv`yB0(Y$D^m9|C, ~[o홧tJHLrb\yd,¨0o2wpuφp*5qt}{닞oS8ʇF&h1iGf+fDC]5cE2PDJΣI^` ̖9nݕj_˰q许dH #PޛҴwݯ S&q KԺ"*g]T@D`QRi?4!p5@ݖ͍uh*x{l6bpp5wUmx,e"pvgGGR34%Zl9;Rg[Oͺw+t8FnKֶ"U\}v򰩶`ZXW@h.* Ad>@>#'j)S>ٝBV7RHCf;V%;SpDuqVf9mm(4eѽ۶6/A8 &\}U38߾EA.Kw-%&TNON ru{?Z]/J;jQ+._=cz"˄H"eBګZю$>'m jZb[l)efTh0` 9vV%ٺmj}|aV֊fg ~.g2*=5 W Fe9vخq<ˏy``80Jȏ3WMߎBˊjA꟯}wQ<`Uu-.Te`qu=ҕ6i =.nә{J%YԮrư0ݽ=֤ ;<]Zѝ Τu]zl޽a`C昚fJ {l9Yw8l- StwRi՝=K F$g;=K|0"C[E.7jvԲ?QFG!=̱2=*j&-#;?GAo:oIH!_5+}mLazla.c)iCXsm Cb(= (5$p~f7ķRs  -uRcuKHg]N4_8}X)o~b1Cj[}I~Tese#:DU]?Q@V?"_Q6I;0nV/PI*TL.S ?}}vB폇G?E[=Zy}:X?Jhdv| `}mJ,C gM A z)/4@`ҾPF"4Bu@$௘% pkĦwLc|9>xCP42u = fZ7Yði@һ JMX+`W@A҃ p["Bh1{In _K4m|^8PW`ݓ\ %A(_ߵ-۲p Pӗڸ v{  hg0Uxb5ҕ˲dψ {xq0*aQA&5h`UF꩟_=bv[w}m_eYJ i^@8:?!ғGN4*aBpv0 DZ6M}ԱqD eFqHnaj-K`0M4rYl*qxryyr t\[#(2cE%ّg, Qaoȩ'j2nR5mfZ|*-&w.a0XGp?2?0-;O/ZΊ3-8k%imYf%l.ehiOt66"= %Ⴑ..nIbb`)]ÅuamC0oQQ[k:[5V kֶ.X~;짋|~gḴEO a# Fy5@Qa4-[dz/k1 b4O8<\waY+ٚ\BZߋ ]{ohPoK܃wUo>-zĎ3(6wӑo{1M`HQ0vt`LE)uƎ ;륎X&[f:܌t1%JCx0&Ӻw5OXBݷiOrY݊eR"|wh{z_O 77< AdKYE(5"M,`E6 G(nT=$ANM8C; ' pɚ/BZh8WaPMEn }|{ZMZx;S?WЗA”G!CgýշƷQal)eUȶ=+440C?¾T3 {6s:[)B7'!Xɶ*n+^쭶57L0ay T J<@o;IYoe^dSt]ɧ&7`[o۬Hާ$Pgб^.uӣ.sXP?{aER?e (J>XM/\i4i|~!e< JW;Jw>OPtAךҦtFlUoҝuawj}mwҺlUZlwYDe{Jv^AP{wA j)Ϊ: p~qEU]5luͮl"x)I("CU90<}i_Y C-S2M4[芠nѮz6j ,!.jO`2KtD0ުt~9wm%6>ы.%:֘xNF(Cв |#4ƨ#HU}k(O$a 3$O ސ'[^gfC?^m X9(Ot-&uUXy*e)Sw뵶.o1㢌Sl "Zd:|!-.fYS/ݤӴ&&W-۵N371C+B{̆SfuSL߾"q5h,NiU}rK4YD3[Pjg`1N<'Q6/fL+Sm lh"@;mgޭ^ I>@#;E0Uh'ḡlJ HiajQ=ȞI[ABzFz`:*,g ,cq5M\n \8zw!{=HuUm/iꕅYyQUbuki'g\(OfR-0J\I׶lS"<)d&"`Z88d/=8l^27٫)Ql}P^gH{',N-DR$|! _*(6 sgHOrS,bBNSPq["Xk&ቶ@ ؇9,4fH$C5hSkm`F@ȞC>*hƒ1)"`E]k4Gr`JaR)-TEtU,+m  3`@tQ ChTJKLBr;Lp-~;Y(6l:pMgPDElD$.n 뗥ґ   xK>p}8>OUCõ>rgh84Kš L $SR{uSx"lj/f/Lb cA2%% -_tP&'@yyFa"DaK]ǸLueC.g=Wl"9b-s]D.1#l"@EGaj# Ao5W#1oMr`Rge\IM1$9[L-%+=,v,\35m$(U+t:O~u*ӡ\w$%Hhԗ39ɬD4U) zD7f(,I6}̸=崎'DTnɁu2~kyl(suZ,m^OBZ [*T.DpʒZQ7Sp2񘈿\>*f3E9̘Ff:45ZJ'EmZKK9 կ* j~J y5^L&z\gz񬥓0IGk"J1떹C=Ĝj#h_vi+}[֚]'47+)iE{g)6ɼ1=20*ePZj穻{ۅШyV\`> 02yb{~ʴ(QCҒ$® =2 Ȗl=z iuAZ=#G=8߅P3}]>?wP{zAyF~K>wUYmknP00}FnjW>~>0>>~FК?ᗟK-ä뿯4z&|7>U1i*~k~w?aoa<)6˛v?ORM~6wU71-uR3Ō3JBfmib =w_3{cW2IC{?:lh:`h*kb1킝sCpkPQ "/^UE|Ip0֐zGI*9h<61ѩ_vMc^am_$wkj:Nus<N/^I,PJ%IqP.h㤩1H}kl}C?خjKT25W y1bWbхY{@gfiIARAt_F8נ{nwH6gЕc;_(1k'rf=(ٵrl,ǢuF V=AöR.GJC(^.j(z$Q<M_끠vϬ9(ȕ3`%zhL7ZP}f]άUc\ mqg/Vd$nI` xvlMʒ|e+[6~={b2(ܽ։8z&(jQx%r!fFZA}X_jP@͛z&aox!Hj-vgxszQߋn R0uw8#F|^ ZHBU&ЊQ@{m:rrl[V lQ:Ą3~P$ُp/ ՉQ|SʗKS,NMK҅QyhJHt^F.8`<3 jo ;VP%5Hϰ_brQɾyƊ^SBEG#GFԩk5:eе`ofwNIFm$Q$uU"u 0ڄaY7}Ʃ18ZSPE[tvkGBMے=pP-粎wa- zM2HIph@JL)S]oWi˗QN.*{QARDyN*]5Kq*-0 (Ps*bqhUCf$]JSJV3Z\WIkTd[ "O>YQ>F% 8S$g]pck YeghpeR(CtcM<=ш8A^$H^hg" 5YzyjsE]`ܩ兮]c3D!cƴa`Hz6lvQ @J8˖5k jGce Bx7s@ (U+.ZPFi4Ƶ\P1:;U+3kQ ]2ZG,YMDKsMeuΎ\*8VV_ǨCQB'MMe$uγZ\,j筇*hfj2Q9Cl"H: 3oh[Ʊdfl I3$FWMhƘ螙L01]=U[B)Tv/m8!Pz9v~UQCk D$HqE!):G k>vir%49#h!;#v c$* *p(Ā.Eҙ)=_wQ ePH>ڢoeK]iI]^:9s)c3!7Lp,H~Ը5][Q̺-㢥o9YO] 9GV !bS`r4sBל Cօfk|k ,j΁@h:?hhjm!J, jlBgVP!*JD."$Ha֛f}`4 ]hM|v6{m1!b9:iu(:zpB)ңT&}upAZ:j[n/Шŷ6=!CK B`0F$#%3S#MR$uqPX@XZ 'a+ɚ8fMt|Z[3c-x]p"p( *IhU)=mSMH)sX-hs*\;IJv0c 5)4Y=to$ܡWY6͍\7j':7Mx8RҕKF;ߴ6& ʕzڤk8.Vᦹs!(UMbQ=mkRc osao7ڰ"&fm(ק&أb8g-FdJ00Jiu$(3[G)p.>qc\5:tQ4ٗN, ep&~}wqǥ3kG59-$fq\ꂣZF8g@%,#(k|QBR*b7Gy^?ʂV4bg skUY=uQE%৸C7F"X]4%B+|I[6*#N$=Sv*vqpRO4D8' T"Tnl/O}dG_ &ŐĎD6΂{B, op3 %jlXL[Z̲AW,+mI=n0ٮ%j@-aq[n:mJrIS,1(3fSk vM9QӆITZ.8tz2] L ZΥ;؝4(m%g[Խ]pjQVRLro.\xb@ЕFtIY3Z/KE/X2 B^CJ,T[Mtq}$].pFvr~p#I9i\%km??L[:J#g-[O.S]M2t_.8fI7ĮΥc k1%7Qrofj%s7(cT;zLblT L)\I.8FgHRm1,g qGgEv} t.DE*f:~ԜR` L/uDkvcSZv)K)fMp 8xSKa?o Y-x[iPsjc8SrȬ¿-o~QYR.1/j6lօIҫ`~U~պ@Ƹ ZY޳\jOɪR'c8Ow asڝL9uiz}$_wO~dvYQxž_z{;NR97.y@+c.DT`rsX}E 4 DOYuQE5ϲbrӥlև ~N.}psk]O7]**֘6aZ"n['|? E1 a(>:vjٙPcIvd;V- ΟŗW0?EG}?1l؋# %>@T\3`X~P}ظ@=/S~Pؘ om~pR7#7&>NPӉi{5>K!< k7?MPTFh^S$tǩQmcT629~}?x'0 oFyA#aQH2zWX`Y/H G LOɆc<v "NYjf%1x=TJS1F_d )aWEsCer 9jY{:[qu{^:ş>Z[2S|셷7oFم3{oUuݛ⦛ pwora¸/Jӈ#v%wovUcGETG;o*Pn (5>~7=S;b4*GG (=_.+Byk|4v|M,F~6d0Z3n`d#{Qu9Tn4}x?'/=.F$s%b@b~_谻mx5[h?ەm?>fgh`=^hD9ڊE#Zyq(/ ݬ(|V^QۮTyMV}#"5Vv7`fKdQov>qcha*g= 4鈩~}~ qZhK#R^8k`d#Q=-pe)e:؏KR! s6{7Ibf`[72ݖPn 9F#0#+!Tżnr_k콯o7czk:.}Y=V"U!dQBp.iIN'JZ>*,=|;cm@Lh_HUz@ [=z@J2ia$؎Zo}tJB*4 *ݠkSA1.[$ !ՠD ~w͌f7m+5fq-mƲECp-' EmHiI5/Q~R ?EAVI"EҠ1j4hҦ[iX$0WSF+h zޔ>:TzR:bI;޿2IeW# ۣob`<:ƨlTNYՂ]CA 9G2gә\^#zzX-MV$y7YU9i ,Q'<6N7<>d_/~WԺpBA9#s,P)Q9V9eН;-U9uMV};.k<6$AdƮŃDVgO`9 +ý:,!(RNcҒ K5x/K@HdO]t>~ 3-VMk]moR,QJz"3iդl"1`,q&U 6|Hڰ QKc9:kt/-:E.++^|<ݴs߯YWи^?ޔͻWQf8:'%sF@Lf}U |/:aW>Q— P bw>7%{^x}C@{3^b1Cd֒/rK+ kck?SI tE:,Yں_1crWlu7Hx+yq~Mf<WzV2zr7YdErݓ/8ݟT V*kpb¹M7EjM|$Wf~-?|d'e0|V82酔#@~ƍe{.=,E'xm'Oξ~]U2wX ov~B}[ߙ<\Ib B8Cΐ1p0ʻʕCagRև*~.[ τlQumWo;UwxYl-Jj/ZEdk/9֙\Z"΃%FoqLRK!04[FW5증7l})۴ek_2.T7a ..S['-WtX. B&sI+V5?Z2$F2 \ !r"; `9L2)F* # Ϝ)(8 МF+KFӎIFKIU ߕW5ەWsĎ۴g?LY{;l/;@&v~YRЃ^I.@LԥEK?,7(=hXeƇo>;W%4 ♨j[jW\c钟g=ĄㇹdLLde Xgvp]-H{spg0(-0(19ۚ@G,ߎoB{97z~ + v`h]Y?w< 2K<ϋjVU)XoG>]>@]T5~jj¼rD dRg{ pgȟ%)wO!ONQJz:+ʌ ":trE4'O?:N}nA<6yw0 avK!g8O<{wSFc`pӷWٗ߿1e&_.]pA7xC=WN`•q_MLw M>÷, |MBy [sj%j#Sf01r=†z H1Q*0FB=?I9ղUjpol<˦WA:h.Ք)u]~j+=q ߝ`~\l`sel|ث`jIޚ0)˵e$>CtC,>*ޚ$^+l4ڴn) PA|Tywr<9{8NMbAsbs$srqDRx#6Iɹᜀ`*?ћQחU"M5F&؁Y{"y2fO(z+z>0~$ hnC<- .hKAY9cGv t]΀& ZF T{C˜ K!@3kc4- w,(¹ P~/~v^Q<l3\i( fg5e`auK]"uJ*jЈ^M3#|Ts!'L'_V'./#Dm?Z[UHIBEZl$:_Z1\qǍ1z0#0\,+#g\`UX I @0`r#9 1xcjfz}16 e--R+Ft:yݦ `aKm>k%#s#!$2H1V[]` 6̈ls)ۼk j?0jQ-R G"QK՟'B@|+QF m. aDΝ&^)EB[t0I(ϕRT 15Bc,]k(-|,uM2O-Rǖ~Z[+ P3ZRn/d[Ib!등Ȱ"P2r$[c\ _|@IQGDL>ނ ,[k%1n4+%X‚rsUX}'@ZHT>둩7 Kα6H5Ļ&16FO !Oi D{R$"7pʉwJiT`3.dN4,$xz'&4Ra6/4^O# v81ZH);h|M aO( ER=)I1D5MȧcRKe1KĤvUFy}Ru&L|q[V]-G[6p: ni¸Kڠ܎PmN@Bf^03J pIwL@K+$x/?|W[NqzYE{ag?XvyMiF'ĭ a[v\Z>RMa9y: ^jAP`:\λGgS5ARvqP Yw]EWRqq! SFO)%%mpi=3MkJka ۱u`}uӠFӚpUTVD+{*CK 6zTxtw].6 zpSXV+$"c0qq};8f2ڻRhu[I7֫0fKn8sn2gaMzXd0Z`|0yrPvMs,fl}`_DrXJT9h[[=ef/r;œ4N&e6To$ˏCۼj aˋwKrk P`hD$ʦs/¹ \Y82.KX|4+`oNU7QUlƀ<v:~wQ842}8sqH9#v.bv_Uܒ9iA;j*wЏ;_yy <˙ O~6=/MaG) >HH8ЊEFJ~'o4\7Dǀ9PJ6,ɋ((laP|Tp 8/~Npgy[ѿkʥ7FzH0KBÃ_yUh+E!l'>E/ٰOl^-eL5K s敷c?v7xMfD$Mze<FA&8Um{xh#-z_MңTV{e*y-yT+TMx 7hGJ⡉@l]nOmނy38TqQxQ#y0PF} p&^47 @~V7 N#ztOu6AWCLKǞޱgol:)5G(S.*@N沞g'Ԯ;4XfHS-| ලyZZJY9E2w=vh7j Kf|%?߿(`TygEiϿ0D7=.>ٻ7[ W.x!qሆ*:&Cx 聻ew@f<]OEXY!~+P?zpO]jf\ .;4e 6w6Eì4erɀZРޠQް\1})N8M [1E)GX3܊qtsD֓vt-`6 XL=5)*fMqlNgNex!qU0r4uHw MfIDK9oPd[Ij: -Cv1E>ca50¦.6+äg^-US+u*=j>ZlyzMݰ:hGȞG17%`!1|@EBEsL>';yIFu`LdgLMxWL WA[Tl>q3A?t9ehN'6PVL:CJ' UZ b~6qѐ=w`}wZrrv MiY }\ShN?) .߿z_2Lc_Me24dkPOД8M,8 3:쮿j|;dpa2ofgvј_=慘YJmc?ҭC{d_> ;zbR˸?쮲`1fY<ތҋ9C>reDo&Qf. /QX"z_ġRdMڣz$4:@[\ȥJ9rOsΓEtbQyxs-( h9s-JM-˕ ).ﻗe>*;os3E#nTtxf6EX*| OQa ^#(6GC=\A-?0wlvc$qJ<E.֒35k or+?V7)Zt$70v$)C TlSaShQg{ҘY'l 'e$ꩯPq5QYy{\kD0SSJ8kkҘ M,IDM.Mw~+&_[ӳY)k3H}t#!bLn7ղ帧Gg.SuǷ,?\˚!̦i6KsgS٨IK*JtsyR~ hw/u \KUffwBzkya~ԗ 7/10ڵ`gV5+Bv㜮.aOV.*[R(7+tziEꍱMVڱ X ">7wa⾻OķY;2lj1T)-bSi)Md~N 娳̦J8͓燘D%+yr0?ZYMN #p 2vY~y<6+_d>r7.hbsgA~~ƅug4Trn_1sO6}6?'_Q ߅ )V/T MZVtV[z/w_%pcYrXwlYÅfvzɆfM+pC󛯃l*Qw !$ácB)7jJ9OSERdqRʦ$I$H!Mə4d8qSG^=>&@ѮQ )s:3sH9!]琮sH9$c\gjշD7gj>€+$yC ߼Ծ졠ӰLtaR*sL9”7SDcKm䆘AKmR=/u6ejK1v6yU6Ʒ[`ePsZb xM{(yM|5FmgWl+&.e=+O>>YD`30C'p͒W)7n#+p9\r%F(3#?~$CRPCUuu׈+fgBar"DJX~輦V}9EBwMp=E8)V$wCN`[l [!]%);1//9vs͈;.ޅd~t[G>{ø4VZ7_ GˋyT>iI Ttdma).Pe1Cf*T5Gٯ2yc2^h9?N^͍)?=I7/b PyBH3^H#m꾓"@O3 PPLd)fX2XV 6֥`f!P~4?yWw=ŷ3e_dlw?n(Ng#v7a-u= zyD%`5BɔkΰNQ5;,/)zf<'StzR:=)NOJ'֞NOJ'U1=)9=)lItzR:=)NOJ'/Ę!DN&5 MBrS!THnpƍS! MBrSɒTHn*$E!)S M5ڦmjM5ڦmS1B5}Ve!$sjq{%@;THg8 G^k XPnń gH9>'E^Y2ԒOMf:u{ͳʷ)]>ȿ'f BrJ\~Jίs&V(upJ~qsիr_1G6X-`dL~b%&h0u-A*kYv9W s%׹ V*ma /1kЫ¼˴xU3z~Qf_א0XS|Ǣ)Re۷xŷ?tbLiae4Wq|xī5 bҥ_&F@Z `V|`Kr@ْ!\㫂38vZn6 Lbg`Zaڐl?@*/^W77ӏiS4 -Z AnòD)vٲ:_CB-1Wk#*\ryS9̓TS;eԽA H❿Nr@$^兩j J(H&ŁկsΛU(~E!P "BH⽔!xNr?|wMzv>s ?3rR1 IF!G<Qma .AcYGy=Y:|s&ᘡS$+)ckbyy?ڲ68K:*}{in<#H`mw"--FVD"/`;Xd^,~K@9_w+N8 2^2b9@JXMӔ@(4D$X!Ԫ*|횷9-[#[ {^@e!*YG,Q HĈ`SR8iDYí"Z\ `(W{Gu F6Eׁ!&cGY@c416p"ra煔W,0Y]0XP~Vs@%kk C~aԪ8j5 7.>Mx9}7El >j N@\QӶ^ݜ4Eܵ9 7#!VQ2ߪiM+9 TJpЀQt`fynQGГ*5Q;<.'q^wn ~!Bqv hiy?Krw2YRR4Y\K=A Tpiy6:fƕ]\vIbt,Eq Bk"mfe9^p׺ྪPQ`H&"-E2'0| mwfrN{2Ypzs2EEmc%g(y@6,6kɋ|ULcĨE2o7ۊM^0EiකJ?{; Y,q63vVlLv}7۽-kfB%Kh#՟|Ȱ=|U_o2z?’}$X$3""#srzl)x$>< _QK3^YĹnj/W]OZv4c͚wi<^d*#\UDRjg#pxlbK 'VCٌ>e{H4an(ݓ3N{E"G\#nRv})R.M9 hi""v#N6SlۜޛE%;Y0cNFZGRt@(R Ccrhm<򞛈R9&^(sq rAdVY]ϵ3kgУA+YA)heO5P%i)4J#cEJ-Q)ê1-G[Kdd9rdA|y$nɆ x8%ӌHn>bƣ{>{įF4C{$*zd39q3F31,;q?}:'%_:~ SHOZ)5kJ^x'2h{'\"V4RT:Mem`ġFZp<= eA1F|Rdꎍ0CU5:YĹj/jHbrj-$T(] R\ MBs:ؼ{os{n!E4¤uR@C3f"cu<,:/p80HZ&&(s8怎)93hY}Y٘&p!W@z5F Xeiv:d"XIQf` J#V2T4mÜ,=b惭o_s)C޵D"9XX)1R>]Jesd>w:\FXLչs,fcwbP8Zdt;j a]`bGCuMԌ6rYP./K&-J8>h}WWqY; mT U+g^9f1|%W-l8סxz\q $63v@[ T!#>C OXшkOv X1gph5sڶGޗGOC?)Bާ9cWtwYQN'$PzhuwQG0_&PKhnR^9~9j:ͧ[uja~ZH q&\_ˆ>4t{_.K?\P?>0ݿ]VTL92z}bݽ}Q;U.R9+ X`L9t "x7+> BkǨZcL7է{ &vܿ,\NPf{~d SOs5 &'b+aE<|_Oi˷;vL~eB.#昘fjwL-팙ʐ.F:yCNfzg "4Y7DΉϗ&Upx^Dx?O bUoJjVzJV# QgW7:kucuARAp BN D&nN8s㛱:pm(vԕmi rg)PR'm?s)Sd%'$F 64pqwxz.q#uck!vP׋SRb\jH)^U25:H{z{.E"D,Qe2MK1I?P[ϸ\#@J4Q3#-6#ց7m TLpR; yߵ"fmn#"sjΌiX":8?!Ǫٚ0P~y HF Whϟn}!=h҂([?_>t6.xxe$&4ݸHk`s@7R Ct?͋"fLY(} =^Эv A,iMqVTÆEu`V[9!*# 7SG4DV'”NQmi]Ƈ1T,µ2+/*Z Xe;n7sptj9߼74 g g;fݥ N&036_~/| c 6 ڎ[nu _x*ڂIv9Qj}("@+1Q"ߊd;M8 ]!Vt4(H@ȨT^Ux;,[lOq4(4mWGlT3q-rd~Ik&~t/?IƐժUZ<۷ (-xDEItVx$xAܖBؼ~q-fX%JV\_ts#~5g@ơL_"܄8QaPDoyD[' f.QrըpTC BX`d@9:dT P+սeJdT\%tA3c{\Y>ȏ }]C ԅm-YNΛf~ hߔj ,j..HU ^jdRk#zn 5o.`V];A N*օR9d*U٢ UhItŎw+3&,Qq. .h } )2;˷<\*~Z E>V\*`KStn-|x\}F5$+{Ur~ʾjstgo ).c$"A(' 5zn]jCc 8\*"x/Q#UyQ97ƠGafׄ}>׊k`>!tv0$ń7ڋ׭h2j-I/_;ll%YHub@O DJ>4rjHT]_I$ ļ 8x}oE>V3Eٓ%iC\E)Dؙ_QQP#(08F]aTM8>#ٌ> T}􀌊k*20TcYzu"Ʉ@1Nv8g8^3(wÉ$2X5+=a7ZgH\ւ+K`D\ .!2>K`.R!r' bc0Pؤ2*&FLW(d@Fމ1v-Զ>ӮUV|Y^qoT\Q whu/fuxTZ&Mx 1;NK,*361u"-m.#qv<| QrN@1QTI̻XC*FFPTON)V0;žoIbӥasj亙1 *.('-t姿^1cl**SaT}~Ѱ|6&kiaf- pβGqSN x$z!n BcMS_U` H 6]]~dTdTDq˗`*]?ϿBv% vx~%^d<TXd! [tQ,8 Γ]/d5M1D97$Un)̚хVBW4F8.lKTڕv*k Z 3#a.? C>Vl!rEMXPsX9}?u~knhR+^q2/t*4!NX"o-DQqVӂ78m\g͐*bW[f94Vx MT46&XaQ/P#3{Vq5a펫 [~e4v@ >pB7)E|(@걧f;iby-W~]q_E[K=9W/ϕw1))" RmSAg _tPuZlg~/S.Kv=VF\*ze|H|-KzMh9چKiȀjcD%n;9A.T;f _G䥾KÖjoG~K{CxY=Y%)J<#DdCT)ZrI*ȮևI?|y ]wQ᥏ZP%DqyNQ 49T%(Ik9xK:]yYZ7(0$ٮW}+qkgJ]>|0 ؖpXAt8y|;gD k@]ay@Sn#gbDΚq1Jf2]H +ōw4)nWs/~M*? N|ݴmT&:UenzVBRaUfBz c#S(ʔ({2 r}%dv3rKc OaT\{ィR-N@5P*m$mba+$sZ (on`//ဌfLpz! Qq!ɓ Lw{7ddm(e8K 1a3EoI6A?f32 +~Iq\@ƟWR'% "r|1Df, QwQdbs5{)̫t H͕m]Aa (1ú9m!K1R2 ?]FaV-kA2*GһTXNK~Jh[ ^n0Z±bmLAcݬ9AM t{@a2_ϐfwmC{KL#[:Muޞg:kr5juWӣU;4Bcϋޥsv,YS/_Tgy9pmᲊLJoPWm7 ?fmHxȯ~É$J=N_<5Rd\ap)P4y^̗kϝ^6I'6cȏᵰa(B~ÓpSY/2r}h-gRo'Kp,8&Wn-Q=DA>AXDM`VB$ J"cjI㝱 /_8T%2 &v3T؉ TluԻ}#o'K0E.m$FmFz:~LfwX`POyrL%Fy"Hs酓 ,r/O`^"_%B*X1 IDBnx[y 76&H`DfkBWC *tE pp;ȽN^S:@NdZjW4p"h7cėaEl n4:Q*r=C- J9ᕩ;U]+bD֭HpjǤ[KKba$שq%HEV 9o81W dݝ@}QG(V7P*[XdVAoDMC5܀ aN Uɟ?*J%k[j>u\0qMA^'}[5Cay$+6V{grbWSHerZ χ%'ːV:s2M9&x>?Iw2URw/9?̿].}JPQPV6C!9 |-q-ѭo$Nl!ɗWLfAWN}~и$̒K:IER)nqūEW;tsA6Nknn _m F6Yj]tx<]DOOj{G TtB bVC4^((9|j;kX.ch$:5&0qAމ o /-"fb+- g'<iq$)&tym׫_v5FhwB ]41G#S,Hr7Ҵ7QሿbssY~C񨩎ǹL(Cd._P&:}]eW[v!HqE/j_B{1I랃*8 ќeDȅwЮZ :%M*hhY eYL uJQ!%ލXebW[`5' " AgEU2be#S!Z!'E<<YV LA΄upSPAo3L0Y}'kO'ٱX ԑbEBH4~s :j#7kwn1Ƕ9XXxm=u2RAm_x/<SQ }|❊ N Q;h5 R]U^$ɝd J'3 5h~ .z9N1BWٵv~ۣ mQ7ʯE%T2EbX}nM8B8dAh޽'WdSh5h6Ǖ- 8v [Qxpf l8xMZfSwV b|1hd]By^`嘖u@ldĶ_G|tXMgUS~m6(C @j^4`ZNrݢxz^F2Ì*KS4ʹJ2GN6Se O7;`pKO7#1ycCsaXo[!;-7gF"LԦ;EX*iv3QF &,3gzkI~/?چFuX@jmӒ xNQGE~RLX>x ~ ~g v̑izVQ5S UZW&/O"qoG^%wM;1(! :zn2Fqy*QYjy6+L"O4a O5|I (_-Q0xII(58bZW9S3<-q$*#\RHά7MRaʨgY=Q*RKzFKK; d!QNL8 $9fOY`Nh w8IEtGHlfJ'"xF" *dqn嶰8bCw9&w6n7> xMv XME2FJxHwY) 3,J)׫Dd'\`p:*rE&CyznQnPjeSjNqTf[ڍRR ѥ%Yzx-Ưavg»(x" OL(qLɹ4S(˕GUD4F 9S*ˊ +,N`{}WovS Ĥ<}mrqeiZ'"Y?/U2K&U.j a],tku&̓Ь뇇\L.O'ivEɊ~[x"tWS<|:6QhqRZ"Ce5tRBov5nWVJ2FMppĶ4O^s\+荠EB{¾L⤘dvo(tFeuU4=ܕ=2F0%81Z6t&dDI{EyxupjT"ZK HXmF ByZ*{|;쪄Z| ]daBhS3tR]5F D [P`Έj+U^_)o>PI ~|t^dp /ЈdbeEf(#G;a3;yyt&$ KnҔSwGpԌqzc6Adž=Z'NP]}$m[r:On$ȋ/o/|_r{?ErXyx&武Jp$ zULkc2Bպ.7 ?[93rU=YGSUTȇT;Tgq-oaRTx/!8Mh?SA *Ia'en1ޅwSw_s"0~^y 6d~H{:T]wz9<܈}۳h z?f9IL~;RX"wn]wQ%c+{W,Lp+8muۤXbz&3槻+[tb4iIF2y 2f1)|dUZ!Q]v~hߔ?C֓n፬'XnaeA]Rv[kGRO ԫqNΟ'}DklFrCjwlFn.Sd*2FHdl^ EQ,qk3s5v|'oֽޱehƁE5H(An d6 5TBYHƑ+@\MJk v GmCxЍZ;{:%lGT>mp1ݐ F`٧b^VBG{&c&iH>/D;VQ"b}!HNM]AGa7hwy1~\phbQ4RX%2T`w9؆& v,f ksocJ ~qby釐&*YD!!&-1ΐ9F T8!+%KƩısA ;Nj79ǭ fQʼ y-C+s^_xۘnL-bAF#)s&D9cPhfݙ=!+&"OE>|k%sx:^ Gȸ QͲE9fvBF4}HB|i̕>T8x*HeN0pD"xD #-aWoh9t :ē֨[g\ܛMNMLPlsq*EpXOCM7)Fmpw)TL-uHU =\EKsS6HoӋrWT!$,o_L_jbHV&C-\ )*3#GH -m"d-zfmaڢĄHi'C/gjwbϦoY `4H|g@+#|FR1qT{0@=PBG/&c #ULO)'1)"Trą8񞏒y3Mw2ߝ/MJxCYț;ȁYCDy\EZ2K]̳J4Obzlx{MYx{/0v4wݨ""(i[:OZ杔Oy>+Ļ|M'-oچZ'o4rNJ0 k uH*Ԯ.噑jSڔ"\2)7V4: e[!t& G7k79!naܭw[n4ձLjTpN;?V yC34Zp]ʄ7>u_w`>z]7 *=kKɃf!G*fnϜ75?Ù)10skZ.ͅfP`"7@\t;0g' up5![:IMJKgo 㛛mqs909D:w\6kxăo^.gc7?D Cڃ31>R0o~GNr婼c;z=.Sb0|nunxsB p!QUvEP ψ{Z̧lr }{a7>cכևcn|t!-7Z]d"4x^|`v_⌂34Cyڼ$#"}3둹ٮn`#e`~~#CwKnd,`\E|,x$T-;y|ҭCw/)̻t& t,HAO00^ g5 #LvARt sf ^B#4X;98s`03p'Z/wǙ3!(F6Ž $A<܇4FfN\@t3-j='~<ŏ ,䥦|')5%p(BJ^#3#å弰yOcz!9Jl˦DPt},x>_w-Ɯ ZdB!J:XH=4FfPE2զv:2(?nLd?ScB.S#3<duǾquU,v ~lK~A 5"pJ }@ֹR"M8(RwOcdH!ݿ$=)QI`€ #pXQ&xgȢ|n݀I𳁼) ۪[NUh-IA ΀ՍSLG\qLW`'235Ѱ@SBE!4shl@N:YlPkX9bQ4SMٖ|A~⩖D>xftnq1o3.?}d3̳jsS >#[# e$=4FfD|# L~Mq("p[Q7#D_TRH=4Ff+T;ԱĻ /afQuK~d*;L;j`($#VwS!u2B˟k{h̜ +lcUÐ$+̗ {/P 4^86UGIa2-`/0.I.4.ӱMazQjrG~~\Z}p9$/~q3}J}y,?4G_g&66"60 UR@tȶŜ2piqBxMAUddpxOP ^]$Q0aH2ϲ&Ǚݎ3<dQSݬek>l>"cG晸3E(@UjM=4Ff̬WUAy0FVȕ\"T KɃP1:V 1`<$+qڗc&$ɧ * . \1kipNS c!o B:}):S>I9c~ƺsk;KMI.$NRzoԸOp;ԧ~p {# " Gp{}+㭷RLJ1q1q\-nir'KW- XMj3mqkJ߭&q+lB[R.]?[J?rʹ meP{WR`#òwgɛdz+zn0NIrI'$ItвNf\v݇-yw_> %g4`[(oj0S}ᄀWǒ[VSqx)5v?KsO~' PBc5 ,EEgv©z7'v;F;RmOgjHTg!Ҧ/Ĥg{h^`[~ٽ\*7 nRf' Q(CIļ`e{w@GrxD[k%8U0N"Qt?32O9⧔My$ :l9p)2)[I{ܶF>VkSo/ѝ@Niܻ [fdl:A&݋}D"Lъ0qwp0[̣(,]}Ey「 #lgf̺2VpkfO5)Lv)i (#VnI^MhƿmB"2!< "^o*_v s}M~@f0syWfS6MWDD0L?lO|}񏿚&̥+Xqw\hv7MfY:\~קw-mhY-﫪O|9L҅qLZh~}Y*;_ņx=,5lfo=%oA*~{"`k U cJ#66HFojRz\b1Vv>`N(8𰨫/MլqJV|9?^,G0k9Yx89uX'Q}X*h$f&;|7Podn<;c㖦%hEg٦Y7˦Wޥx;kLAl_ܷ܎GuZFW렘GNC0Q;)_`E3e0[gn֛mgoݲ֬7-2ۜ\w]ɠ:5ܽI2Cdz $vziVƲb,60gGcb#r;}U]o.1gZ2zgMY>lE\/pP[A58٢5CYݬ]&׋M{ڢ$"_hY"4a"?bkt),je+^{kO==o 00L4vhPsGWsh-0s'&)#!;~NJqTQɹf.%+a -J)Wk%G+NIEG#. ([Pu}8#8L}P&.Zc: HkSzGpuAH~ׅzjɗ7m/w@8X ~4 6nW|fx|Q͗?0_^"{c=6Q_0\B^D/hb\^{R*$G^,ÝS85"쳯]S^|oi!kǰ酧:Э #9,VHbDXd*4Rc( Ni6)$K]B u+¥&L@ViǬQFYJy{apȳ^ |!_UGfSaMt5۶{Psvc cH殀8߀_u$vt./V=bvnck28NH[#N8DzC {uɰ9<[ܵi=*K*#R"%cRYǮ)`PqXŠB OZ@q`cb^)ό%%ŷt=P%Z̄ Y\K`4P-ԉW?o{=:!AH\jE" vZf;sH%iJ&H.>c2qţWOb5&Fx)d;”}(]PXvܶcnۣz{Xqe].3w围@nMt͗ _6aQ2@<]~{ЭlMnןUf=tXjO͞%ؐZKZT}`dz'gR*M7g.~u@{plbx;mUWSnV{Ic~o۬=G{3_ƹ;W5.C5[E7:a`3qbl 6W?s2B0FaE[^/Ytqx5hgM}fP ư=3؃SDή٥bf٦C,,t@`1Q:mqkrTp5AsV~ږPw02*nS7ykSuZ_؋ǔZl?hۤ˿vHJ"Vc}~o{.D1EG}0Y[P6kmH@\P~0ܮln _OKgdHʱg֐C=["gz:v<䩪ݘqeW2:q,`09 _a..Eliz^xh3um*zzFU \ܒ2(4HJSEJFԜb(S{/3d ]9)ctgfۨ/I D0^w}Zh="g(ٳ'$KOK3A0ɵ r B($<3`M .RFXp4 !ޔg^sMZ8Ur[ *ja3kFbq*30303030-20f"`&fDLDLDLDLDLDL-30303<͙N303030DLDLWL i&f,303030)!(>oJ6(RdKJSp2a]BtV=y r 8M(rneZGEdQ0A[R.k%!RHDc 6crnl|Kߔ#ócccG 5qF}UJJ8L˴Obw1س ' _OlJa^iGjd^ wE%6ѱfMEE N&x;ft:3g]? 03+DP V3#B`'Dp4 Р0J1 FQYf3 AJy<+E=ƨL8(Z)"xZsU h\"Zb)G^8d(Q25Aw@FϷ{]= M/}q鸋Qpl`EGb1pgBX _HԚ58uC6x R7r9ZGcpX*e]ibR6J!^ڠLiL )&yym]oVғ='[oO.r|!#s v$~hܯol|&brie@ieB ՚R0> &c ĦmI %ȗZrV:`9M=q6=<vPM?6!'?Ŕ(["Ci-OId^^'5~f@eà-jjMŕKiSԆ2xgPPJ`ƿ08*/jQ\ W|Bge*bu kثl(& WK^jQJ' իW2o:'x4bIU?FWCxKW=xiTWdHzC4¤%s4Mt~Q,|TDnh'?Ňo7")QZpNoU=`,@MUM~&1_+RZh7d";V8v*ō_x=tEǥ kWuZ .AG#ħH׍Nm8@w6UnmmpڦM\O݅¿V$maJmvדej8rnyҋPLБ&p< ?M![1Ōb횱ăf{ڛh=Km"Wg4V`1aeMضX推%v1p՜)۴ֵ_p©N)6['_:J^JG P9aޡj]] =z{]L,#{ڜ}:[]uX6:i巳`ϝE{[ ]vl}A™ms_l7YAӉn ۝1L?e'oTWz3ߴ͓*Ѩ)CzŚ}>#3N3SS1ΧW.B~εǃ]Ew>݃AC0 #q-ܥSGd:G% w!mJD5ʺRdEzC3anJ!+ `yTǖX=.1UJnDG*KԨ!Hn[6ky3yH9oFZ[:\밚;I՘~|o|'3Y$%`#nb<֗` %&%}&T8R9]{~+2] L}Z ~]hXu*t>;}8c{X{ +tMJec0l(\|>YՒ\J8LP\V.J/|S%uX߫ 0 E=IM_:{(nu&hV$Iyܞ?5Uy}m{@S$ߴYălx`#98n,#u !c$1" bL[%Fje "F3d(f^cwƅ?&x؋-Ӳ֢{'eVq.B Fl q0טj{uɰ9Z*M:TxGDJNcRY *@I(͸J&TB$=]cMOa0HZkgƒ@-΀D[:kӔ|D%š:A g3B8jE" vZf 6rTqmy!u=w͡+bi0qX` ˏŇI g&)0^pA v-GIѬbz4EGхS"5``)LeׅsN/GkTRKA .>I_zxBS$%Edž_G*8辊V%kU-"*Zh/|#)9?-{{U70o,h.Jg22.0O_E1/RC_0FOaE~V3و4XUG?VQo^C26"qSDR|3^V#"{#$L¥M`.{B*cظTB!Bja ݯ3Q5i0b24T7V%%ӋIṷ{VH%>,uU-`Y?\MRU"a if~QXn;SeW2:qbmFl[e; iCnIfK߂8@@4Zj `*#͝D'az5:P|L&ּY7jBU#w{K:$< -;(.9䤊IZ-i &Y9!gMA?geut ۽y8Oc&C=?UW[QO 睡໓x>M:_mlobp"%3HGi/-tU}:*nt z>W=g3*MRD;EpI >] `"Ui%FzݷgPuP\T`ۇg}}Vfߓ?ٰ-b-vxzǁz*F-R!C38C38C38CXUcG-{H[H[H[Br뾶-Ͳ"ODHpR8CΈGuo{"ȧ.BCYѻ`r)B2$/-E'deGFi\%/Sp8a\U@fl-UJxYr5gsMpл(C4KF5VxK37\ ^R`cs"u43/eJ`baLT EaɑM16%)xB{B>6ʱf#;ퟴKWy`* ]|5^e) qZ"%U$dI-Ve#ǗX-~jq&$"-lj-:M*lUr)颴Jxh w2t[#7g3<*OobpaO8alpu`}]1Mhr\###-GHKg4@K)$ѰHr(XE)#J>#J>#J>#J>#J>#J>#J>'U8k J>b~*#J>#J>#J>#J>38hƙ(|DG|DG|DG||P@7%CX})?㏃_5ӌo_wo`?>oK .3/7Me4hƾ#<Oy:G3=\; ՠ^9A w; P\z9QM^ձuێaSFfHϳ*&\2ڎ.#s{~ˆ]rIfF3n894L_!؞}K{d{Y hsYکj|F[I|Kl/.OE+&paZj{Y>1vhe"c1 1 1 0"CDȲ(W+w?y&ZA8HtkD)@(B(9ѢTˮH,id9km,<r #B5Pk`8ೇ_rh`hoz# 1uɻС ZW(B{OTZ_RR{*kِ:4Ăx1;N}7:fHHHKjE~H;?,oh!"DdΎa.S9ezX -/<(\1% &I*$ORRtBfXvK/hPFZ\H05-YEkMvR@Ddω (;k:wQ i%kfn['Υ$ZEhf^d˔xPHu Yj').g`i, u̘R`eeΐ$j,pLwNYbѰLޥ꘩Wlr,7N# ʭ)M0x/JF}Lb)Z1G+.ޓ#/9c, mJjS*`*~9>|lc@w _?穓@[s )|vuλxjR⼵DJH֓J&F0/_?LHFEZ-ج5*[uU`٪REic(d4*Gnfdu/)ypbv>pآ>Zz\u9R.ёБ1p#y+uVh6@ $ pCc)!"+߇AˎioȒ X rSIpU:4 ɅnS;Z8vX7h>@Py$P>;%2Weu >g<1}']#j 'fX9JtX}Ne"56n3 U}fJU"[d8 hJJU{vhhۛ-!#7_43{x ,,0 9 _۪UZ˩?5[L}`=t3唾C0ʙ0? ozT{]I˔/Fǟa۸>?o|?\TkYޚ2y/?ܼwE[ϣ֗ mUD,^zU2I2$ s2W&8%`ς $ #r;ِ}| /4 _t,)c5!ۦRTP-`Z=**i$<ΓwmcO?nw-Z z7=jw]ϣ&U%1[Ҥ? AE.J8 ApO4ebq˘\װlKfƘ7Ԛ Yj˞e&1#qǔɅ&*ߪܱ +R L`Jfh?/^2y0lI&OU.K>`ݲ.hy3Y}_:uJy *)\. ZURO$7zf+>nFabڷ9Seқek|G X 1>FV皻D#Y󌛩y@xay):2L_>ͅ7Mtnn\Hݽ7i`pNMٚg<3x7-zڱ<7/A85UؼG[Má`H&xݦL?"އ˫>*HYT\wX*YZL"#u$Hኇ(y0@e?=uRHtx|(gs-I$)cYdӥ0%HV:ʴY#JBnT[3@Z~OZ~hj4nkYO~Tföbhޑy(„"L(„"L(„"LHŏ"L(„"L("L(„"L(„"L(0"LP EP EP GQ}b09@r͎3JiQh$d9&⌒$Z>9.B<:pG'?L m!ܓL|<4{r h zp+EHu$H失lۘKa'/ӄ[+4'TvBiPwtT=AFP!]C;q ;v>{8UCq`ho ,Ha[>.3c`)9EZW(B{OTn&_RR{Ϫtِ:b4ĂV1;N}7:fH!8zA[+^~krI}D+CLhQ$9EOxl|:ǫQѳzʩ+WUUɣa7<,hmD3,Ej+qe#L^B.ytpo:JrO`MQqopW}6TlW39k~MFWy 7t[0QuTB/4/\)Yk9Wt(#X7}n볦hGît Vޱ}b{⯛^nU?D]4iʨUj6u] =Uٸ =zW UͦgԛOo:emLnbv+}pj؞&)+A=jQ{${lhA4/Q#nĭχZka`F=_h/*.Ylq}ן'CvCI26Kw]{(̒4HV)tP `J*'/ /]+d:9:Pq*74+_tR""( ɚS ׊Փh#J)l$j!$<͗/vdߍwj4k,KP?B:G:?.Q zIupkin 6xo٧I ,8`AIϴOa:\>~~HW!̶dUxB KR^W!F.{7>)Lqǔ“HXD5Ɗ a@4Y<ecl:aRyi=XqBg%QgyPv?v7,D @N  b>ɨ5+ 7h uN|VJAxbsjz_}dK| 5{(|{oȿeۛxx rNJD"`fGԢX|6$} |`L_mh\'Y eR,8.4*Q%*P@u^@mt6ۤkOM lG/UasvRjlخ?@zڡh|P^RgC%)bפۻ-w*dzkV篫ż#̐uS7sOֵTdbڙc=9xO[Cз9|0py/pN$y`YJA%AE RE%H]HkAfoU6qpXK_L>_el=usmiMZxtK;9/|Vmv,BQڒ0FJZyoku7J.>f;w63D~pʞ)s˳ᬄyb~b#&YW?l:B=z'.ߤHX\ÊW\-`JkY6_ AwiZh:$u' ź?ͯSC4 \[NBKVt?WB%G~%}^b t{`ŭ>̎(">"k@ ğ]:h]y?x}̶]EKrLm0Uw?cY$0xꄫ%LƗGSix6 `zDtG9!~8ڮvS /_Gb|5^E-X=888=|%!MˇW;i{ @JNQ\^>]( Qx6f ve1' I̮+]2P )ǡ'Aڵ;f-\[$t 5:r<NYy;it>{{oa .N0۔;~!t.3bX2k%~g]9/.Z4_sa0.36;/5 a dix|gio2.1:9@i9Q&r1sqf~xx>7JX-H[^R0J1rdz*[EkW$H "\]J͋]P%qRdNP,}Њ&s5f;7\=ڷ-`ij>iҚ}q4U_65|\=߻KVh4"ٞeInCЦT1~9<3)O)d~ ֠EÅާ+7d;pZ^!H_>~ =uH:o?+.|P)!zyR2z^p$Ց{ޅx9& a i}d [1B Xi؊\"Ro s6%i}݉Bduk00Ef3[4a ${ЂxE9jTA)4զ-`N. 1FZZ/hU$Ify3B p5U!z jdh˼h@B|Zo '-Duh9 mxv |Zo/y laXj%"ȲhnЂx'ek#OVֹf>m)b6d[1B^JLLicIlmgċTFc=I>ZԗɔyX@893B@#V`똷",sOVEJ ΂:P/RE\t&4Csh-g %[Vh#w=}&[Be2[<># ^G=u4еC W[}FhA<1Xw0o"p/'Yh%F}o %2 E J%36⷏ЄxcDk,gPgRa1&=FhA`-y.IY|DA("|d &E}׿Mi:$M[?ۛFgMXHiӏꗥ+UV/i7``#߷190^FJfu]p+mn3&;6i+:;&0vGK@|G|q GG foV"؂S;G4: 17 h_JYҹT*r@a j4R?Y45 h0tIېemNW \ybOAL/9u?7oA~nz +w|Sڸ& &?ͩeRFHіMa diQO,pD^[ZԎt^*k),&drC"1aAx\:|vPyMuI hq *V9&T* A{OF~ Bi4diFyK sk_m2-L`NÑ5E:Znz@XQ y4 @@l3a\ j |O{4;p9D8p+cn\*J tIwbM BZ)CUYbBΙI"U?+s<_ q'),zu{~i˜Jzn|fteޏ6:;ĒW?Y%ԷضմNѿD=ھ2Ry #`Cz; zT*n; v}oLGm.=9OrbΕM[ܻ.j RMBԒ%&a(4Fuj YU0QKƹt`-lB viYk@#D)"b!(mC[ZݱI)#'6z踱9["] T@^R&jt{ԂH%L boA*+-#"M<É˔΀*%wPy5|T@IrHsg${K[r5P d I q`xxB3̂[N 5F*JC'O /n _(IOF_7y'- "*W3>&EVj*HUy!mMtiPb%8L ^p[`<,>̿`C?,kN2ΪdS:ggOgɩTxs. j3>/,Ḓ߫W[ bI~s8ҝ|0鶎G JJq}lD?xGB Lf|9j21 *&UN*1] @%I'Ŭ(~:RtLp4[1] @2Sy@1RXmuHV4.* 2P 6'M<dbhjLT<蔥%"9N3 kޒ 㵻Be !d1 RBd6jQK''N@H@)Vrp{s^J8]q9 U Oy2xuY{+jAɝ@Ɂ.#^L*˫4#_5F @L]pj1p}TMCFZfTle[lt]`091kUٞ-#YZVjc|ܟE\ky7 axics%VS(g ,6^{ژn\f\߽&M;)<҄xT 1=+V߯Xxw ؆_/UFs%JY0 YϣK N1YƇPEފr8KWGZ6-%6$^Ov~EV͟`ep=πym#2 "{.x9AW5QgM`U)ӠM3`!<g-$LeuTT Д |AMPkQclvU8X2v7$c0J4PlV"H>` Ʉ,-fJ2%hy%mv(.fXUDՈ ϋ/x7ǹ6\)Wd)au4_,nKO]vjw6Nը|9On !4L{LP6F\R)@)FQ$%*} ѨRTFAe?MYɢ+ͦpsd=6Ȋ~sf >K*~ k "&̀xG%9gǓqEA \'!";feHLRE *o(ϸ%(#,[хi#a4 o9 "#1NfB*EbJ-wgP -]R>ʊ +u"}{~lO3[q䚌tZhHrPd@\s `Q,j-cI@ k笳`΂ ]O|p5s)1iv)*^ol3^TL\Z] F\\sG^NĚ%owW\Vww5owr>aY}~Yu Z3mkϳxnEqI]B.^`tB-*%ȻH.*(LΉ0olBCCI˄E;%0ЌDzk94I+0\^P1pGBp Agm!a,`fLD$E4(*bqOdžqҰ9{IX ݢA>$*52 y&7<:0/7H-+p:_wۯ}[\5;-Psٟ-LFKYngs (D2qEvVs *U{|`>h=7Go.?Dn{܌{tqH4<2M.2ϳ+v?-GҨ]?ݦ@9ח.W|N5J_~* LaKhJv d152vkWbL+%3cDѨOF _FCgUbuށX QE2}P*uipdv|e>weYyʮ[}X>P__i<&GeOt2p{t1>4[~"gNLwij׺EԕuByH)Xz?X~20X4XF_q*P;ʴc<O oҪ8$V:r=V܀dph7")GYo\`Q ϠSLm{U\tȊxV\OH̖w(8F룙a2){=zPES峛'^oȣW 5-4{LuBaݧ<,_q2vDj5/Ar24Qz]hRDa+ۨ.Ks]ʞ>Wk+zr?O"@Ns m )B Kf) Le @NcBkkȼ G}QK:f2 .҈yԴC9tnw]km#9 ܇Ȉ~0pA;X>l.k,Q-LQ\W_!5×fDR%ΰW/I>MLis:KFU Mh-B -gu Dn [Kodü^qм+py4hdSG{i8W<@ -e@P+e< ,Ѽ2"h%TIn<ţTSEb+r+k\5/Xjy\-;}` d^ 6 f݇ݮxҴiymP%ϽZ#*ʌWGune;g&7L3_/6{l\`=KUS@{w&'mi^c޷|T|1.oOߟr>ЌQ) ZZ$ qj|Y[.Js3DbKĶ";@?L.-ÊQgcˋ{*uFgk R**ku3`epd3 ՚1w()+gI]-~>8h {LFݍa.B/JD1NFF-iH X!8R 4'2do{ }K/~oc6 ^rCi GT>u4qZ0)EL1VƉ9bW5/H{ۥ 9+FÝ[iRg ̋ߢemrs_4 b*ryYb`BG0MFZcZ|e(ЀA F+]ͶGZgBZ(־ꈢGY6QV7 ~՞M~ id$}ղ _,jeV([Hw&^5]mҸk0pJ&fZ>罥BeĘKY+j-@b'vJD x_VlIAtnS#uLֳ UQ*K({z:P3ﶨm mafIerI 3|t=,tP+5Q'DĄaU HЦ ezeߑ% H*fJBA0E A1`&(kQcD]+q/\,2dJ*HMJS`) > Lgi*BІ58-{I%.%%#!@\3[_B~ړ,:<}}.9kVtkk@ĽeE<X'SAe I~gݸRVݧί<#ڳ weGgL(sl54魯UZK8-HCC'zhN[J i笳8IP𕸔(4LjMeIJ=wϹIq'%y: Wp^,w OXlү'U&x2 ˲1zȕKFJVsRTeiLkS0-NPJX!5RYОd{%"&w< {eӀ20{&U[Xa\ K "&B!8p HbR֥M:i򰔛ˍD~:R4,σR1gg׳(V/V.>MydE/7]~ Kٲ' f.gx™ ݙRYWLNjanB^Ls\أ_0/_A%/o+GȪd=r? k~Ξ샺[NcB)1-ޅ~fOHş 1̍KV (Qïtl%CU7qp;n4߷c9mνC@kuD|fߜ 5he#aPN̈́i9fGO79nr70Z̨9R@x9U@t>d_d >GQ\ '(;uʱx>eNӒ' =~q"<8M7o[*9gV=~w;?>}3 bMg4$wT!3;Gn(ZfFS#q#(M)J H%9T2Dڇq۔p8.CImNJƨ n'=3}-~Z fd٩r~qHGOJ;߼Zmի8_t1WctK(ˡ8WP~SSP]skS:4]!UGlR [,񪅲]-rJSL\ǖ;7?*ԝ}}5ՐwoSɛ wл{1ՐXuq+ߧb́R`bcM_Q'-\ o~OK׌j"-@CCI9؊UV )*_ƹx~%{^v ) Kp+:@=_oY>Eu'c>5]nW2Tz2בYnH!WCne%m`BzR:ZґKW%2kbv T KWMW~=+\L;3O{X>{0y4]-vvWsy'b"2(#F5e6j٤.KF((Nt'RN 㐊&ލcE.&۷WZ@:>Ch!e]nWJZ$}jv3UAG_?jH_ RkYEEk5{^U6vFXe3u'^U!1 ʶLܶe[ѶVt~Z|oA $)8P_L!s E X Nڎ-GP9v8D} zadvTb:):F]WY_C$N1F<"(W"A(iGj@}kTUQhH$Ȉ0Ks!/Τ3%sx=p,9A]`-Xkm:KN>^P̄ĊUBE;;q_+eE}7?ulyj\H A8[/\dA) MDK#NT,O;^x[e 0\Miپz˞ڸ2hec-Z[D܁oڈg(1mg'T7}pq<*Ws庰B_hP^H{BEEtҧz뾃9ߪ:n莪(Eņ]_S+#i~68.%eKuI뤭M$BFcBbk˴-F+{kDrDo~a[5-_ΓohR^*ҼrY(8_Igh\T']/FolҊWw*lʃLPv(եJuqRMB~J^˟Djdbo^Pyo@Y"13 "u;˝^iT BkJn NW H+YW 7~cfj UfЏy9fm"95wGlf78yq^PwE݃mnjY<.64*dȕ;wCY0qݭ$_4l9r=o:hDNdX'4e%nd/qݘ<xZzra;ޗ(hI*gۂ*VFj} Yڹ)[ύ'/-k.nsY0ZPXH4g/\Ow,@,Kpj$^ F'( Dsͼ Ezai9"i@rzb, 2r$^kozOGᩢ`/;G w,wQ %ܘdL@,8*e:$ 7.ޭtn \XQ5&*puJG)-S0O%IA+ !)WH M|xCbB"zqX# @)j Ǽa:\\ p\8Uj+@5˓dbN/c$cRRR  \,hϸ4rmR(2`q2"~aZXNN Ciy0r̎g#1لr3w\qieV3fϮRt帔rFkg*/ΩT&f8Wwƿyqr񪡐nZpr$ ) `\S9e}`V<~%ƪ$  (o}?ٮ۸9 csn zw7?=c=aVxA19|Bu (V/t67eq0|, fs\gq}BdC3BF;2lɟm6;c ^?;N}XTnAPPkx*8I{}n?;r5ܸ?.uُGޚX /E!҂rD|]\`xG=n1F- @BpʅKn.HhS#i7h2Z A!w@Jz0b T~)ͩk;v:=mtV6Ѿl훱I;BbFuePe٢Iexh<1y"$1d !Nr[r$˨"`i*J&ΑhA"2Q88WÅjUI.͔%^ZjU,IJ *gyrrvZ'1oQ1*^~/)4X(Dx=En{F (Pd 6E{AOyƄ Jk5Yʭ%:B霒,'t0%(V{P+LIGY zoxDQ:DwV)eK jI:+l0t8ZQLF{{x6N+dsE 7OGV"!x3͓ =/fIp@1^hcT@0>Q*:ik/"k{CO ABWEZc jYftvA۬`,DXJtFYD8*e E9x's065yt薥CU67[y'ЫDΡ%$tY3Gʚu=DRheB"d4挫Ǽ1\>\vQ{濖a\R[/ds͹ssn͙pX*5^@R,| HHE""DI%&:Lb_T++ϫǚPi}'*w$=y@e_&ub:|T$9U{?%f\wr]k\Zk-5Zu:IZumr]k\Z\ZPnjaqtr%ZSL'4qD/u;Lx>B0=&y|,𹼏{!&md01XxҋXs"/~ÌӴP(DwM| ^֚iK epaڮ5y˪w>hMG4C-GOOHOQ/}.}-t=[e]֮gk׳lzޣ6l Q׳lzv=[])b!zafQ j>j-B -oGu @E*E 21|.`:}`u G:4@ -eJ"B8K{E锨G#,Jrk)vhpO4f=&|0.,8Wezr}_5g饔U^u׶wPDPi-w%u}h4 &h%/_ ui %`x, ;@{¿/Tr@Q&ŻP_ U{tˣRhI@\0P+~@M$*\ZofV;xկ佁Ō_ƽ>r/67}K0k򱣁d[X9SqT|[qΔ4YHhjzˇAoҶ6YDQ@)bIcԛLK()X*7+h;\L}58_"h<̽7 F-s5,lΘ>?822k˦559w]6rtʃioF?chsqf9nn{7B҅}' !('QS6jIebJI, 0'2d}T58|kZxs>Xnʺ5>X|꿻%3ҦYSeP(P -LѲ6MR8Jb*֥QUׂUF[\P֮ ]@W t-iiYMk.6V$DU>7`vy7/e>޼\>޼CY{; ~*9P\x|K]ԷϪlgQ+{eԣ}ZWvEP{`ڪ=4γ%9ap L } {KDy%,wÕLٷqTsI~?@c|ё|8/)XB~M6Lϒ"8hᑥ5NAkMLhf_l/u~]~#e'K(o}FBՌA6nUk0N1&IS#uLֳ UQ*K(b:Q+6yȱKj.I/;\}´-P~UN4'0G3β^J%Q3uBHMLXVmka:+IX>o$SјRP6p(LHP m- !~|L2k7!N:dJ*HMJS`) > Lgi*B> $N)ig-a<E2>E+v x^n!ʈ|șyፔ*( pxN)8TZ-P.$8 2'_u *o)YggiLkS0]fo@%aZ% (.?^wcPȸ0֙׉Me{V}ja݅,d|ORpITYPW\j1[MAu Bgev$@i0#Ԙ3ֆpMGk! FE@@"O[a97:жvi[Es4T J$G2"\DDr)OJ,Wʦʦ6:BH"u+"Ǻ/UgUETVAv-Gau%ve4M.?ßH͚'L5ws}V׳PA,>\Yq95ΜjŸb?ng`3ݜQ?7 y3ѿ:J00U|D&59cP0)Om ƀcF MYT#OW'o͟y1sHJ /Q.[(jCl6U{_L¨+z]k&hBͯPm7 N:fGԠQB9a#rq甇 WUp_.  rzmQ=f)XO] *u r|p3e%`091:tܟ2Ն~[ثxү2|Ӕ;< s)JAl߽ |h\rS_-ow]D͵cO~ bFn4ֻ).%|FZdnY.Xme?wq|bmض@Տl)? A29rC;*jrkE>rJNMn,pm^;!c@3dS=:x9.)!>DfSJ‰HV>ۈNJƨ hT׋BH#LF{dz?ĊCv},RS_e*i4?[Wq6습Y<.4ILXzT+Di8ɷdI ԗ(S ̕'2v]MMߔv8߫Զ'{[\.Sh<"(WBf@FUD@b4>n0K1WLI\"jO\?KNPX0 P'h+>3x _MtqODO.y+*G߁%w=WCўVdvӱ (d_s}snV@~ 10} Bݧhrnw~K@DǼ?@D'ɓ?NNjuo} {c$6Fcb\~q{#7|)_OhOp_`2,\3Z|/m+^zZb¢c7;hsҳ1UaOƶ'^ zZ:/|A_ǭF3.zAt_fG 􊓊)n@IuMzk6Wcv=E>_mLzEOHt]7hk՝_0\%#WV^5wk{N NϺ>_18Rf5p׋1o99|L$c&3}]J8U48!E/9^1Wx3/ZDgD ,]\ Tw9 &b "<]% E{6qvP[kaVsyP.bˀ"C0!B*AY)lZ D}JH&Z i"J̗d?/ZHuOGe˥lry/LyZjU{Zuw׃;*vFfH䱿*`bJRҚTu9:Z2mwjsA6Gg=|;s~~f{[~fzG(}BZ e/`p$yo/@MIX(UoǛ4W2:w )np09sH[G‰uXY<&ArhneטӠot:K ZGK!zƌƁZ 1b"iZ+IWF-L"[+zlj/խ Y*љN9 l풲9 l۩luږZߦwٷiKA4]e5[9eznɗ^0|;6SxdO 3}o|މ)+nUL0i0Br}˸[#g:z<7۪Më~f[|Wŝ׮<, ;$sA0O'[ -r5:` BBv 27S KP;IԋTjװylpWwѶ]qIFϻhdv3\N-QO,jьrz =t{}6H;:2qC ʔqDGXYؤ/<T;CN 諦\#h9@0U #I@.?>?|r9tϞ:%vɘ=ýS܋ЌHzKf $bfp]^gs1a;8Q&<@rFϵ,wJÃ& '?HZԗYMFD.To1F: lp6HF/g)Oc.*51V!SmH<-8+69@ 5L_DJ*٨~ &,^V0#)6gF\daL8Cwdշ.>cPրMv]d20Ti#L /f|ͤ W6*?8X@Yң\ )rhڣaH8?x]p3O0o(Y q28>Hzve S1e/[Z6Ih2`OvD9i $CR5jxT8!,$ɰ[)r]qC 쮬 ^T*N弻B}:B))& [)Q[M*Ffd j0DNd(A^pbu}|c-禪,yX<߼UL&oqrFʮ|4 j-$ TH'Kґ0хT][ ZIXBg `*(S:$ rt-!a#6HIg.A@2FHH5pFBb%#1 )DXҠ8KFbR.di9H i)i-a=ŁՕr:hU] }}_d_ G̟%G/s\}"Q3} 9 { wsr}~>7<_2=Ϡ@Vˑ֐Rsq>-a$gL`X) <Pyŏ1m^1'4FTS&|#.E:Rb&y` Qc(,<0592`bU-(<_bP?JO͖%X=jX PDkY>L}v`dhqP;Ӂ$lZ Btv͕K_}A{x®IJ$⫉lŐf $=#*Zh_9^s&MXdo<0wgk5.5JO`\k "r!+s䒾H+UZdWיIp\_Oe4Zi"n*,.bP>mHoGv{F"rnSwe"U%w&:?[VLSVmqkB:ʥCRMgbWO@ȱ\K.iBTUNl8J`JEQ(ߖ{}"qF5{?p"uE(ъlo XCVq#R^<.nەtV%7.<|yߦO1|9w&ObJєy(|HѮYt(=:d〞 Ym/)*K\=moځCng@2(v@Z 00+@ ՞xxΐ5d@nm-M^ϟh}b7}sS!\7A7ŧlBy6~n̷k&HC`ל^#+ ?e٢j`Cc+D RH-> }lw&.Gń%% a[x 1FCr{;~VFS E]f "rQ1G EέRȰ, Ƹ#aREb]g>mn 'Ԝ\{*塁F7 wK|O9%}w.5Uw7 q. tWDLeG)u|CwI|9/Os)B>ԭU큮kM||FYrЊ )$0t*`byCqCru^=sZtԱOͲCh /X>>yp-|6}s>Jd!u:&WJ!gj{8_!iH]&  lp1~ӔBRvCU_DѤL#Alif8鮪~itwd6/EsKw͠Φ}}5RQxS@v[K;<]Na"#NF!ۇuJh? 4elCwRE)+u{?ւ*dsM^{aBM*,( 6B%e 4YP&v qe~qivhDD0d\b)Ȩ0&c"U_ ^i۩(A9E,$rȢ%JOS)bj7H gׇCPt ck: QEM4F* D[Q$CQ*Х |x)?7;eRd󼼾ڭptB9@%d&a,2VwP+MIkpUdM`(Q!  m}2G@Q YᬓPc\o8:1G3x!CU6I=/%=lk|༷t`5Zzm."ck^ɱ(_g"P+"SkVzƲx^J"*":#jxZdҽztK0la=vt&!ױv1z=Z`oKe h)HHQ*-XyCoxxG))V J̍&&bFXpR+4Y`Yz㦤#JI}JfBxX`q rHG |}g*M,2 R)) @6a`xP9 R-:#Ogٛ@Yqjg2N]t翺^OG}grNWgb~P)1i#%@"(˜I?An{O~LC^SGSd6&%+m#eJo "-O"by_ٛ}eB]g:l"/!UJ(M塅xһBAj(A]h:ayQ>Xk1IX#|,`⽇,D.JBdŸhJ,d/ ^ǎHƔjkBמsl@[K>xp`,hl2 y~1 *KƘi \0ui0xh CS:7䔐}Nt2Bim^gxT%i o2lP@;:v CPM@ǚ))(*ӒEق*'y}}8*p R}Т#75ZgҬTM[(bBusѡ0o֒k(xyy5? C^߫}u«o^_Z- !aOl(#%lE. % UB GHPH:ya^B)Ľ-? \;>ovu5:6'gMQG<5Kl@!!4*g6(UɘM4u}\ʕ8dDC%O&U:}#Hms"yNXNMSE~O; &A<^6^*6`@jc^Y8$9YU!-MYkg(Y<53^hݒ?IIN}{*B{^"MҼx.hpxqx16`U< CbJVlGNbv=x,R#i%Q|6`}@˾ ]@B7S'E)$+Idye%!b2$BB> : 1:VuBeesL^%+"_fNW+c]ؖsi5މEd ;BSzߚ8#>@}~Ѹ- md%[GZU׎8tQ6IhQr) Udl|~!aoW)n1YW%i <@`gt.8d j_=AQGY1_#Czy˲ܺc^,QGoMUTmULH,q A>Z i) J؛Ơ*`Q6',I*I;/] F£6 }A2U؟?<3:1ݿB5Gpp:]_( G()lVk+Pe\_)t P%ۋA؄W&$DAX4d)3\sBR{b"V^jo0^U$YƛZfK G2f)΢lJI-T͉mҘSoˉ6`\vAzƽ2ᬋѤ\j&|TE֡قx3~Gqj#5k)mB"[QO<խMm!jyxvT?z6=SҁY>e oWm>"C76JR\ iyY{ߑEw< nY1s.?1<CzW/gry:ugBlkHjNYN?&[dl&Be/80LěmZu6rzpd(oE7'Dxvj@\ɹJ NdBdʛb3?[`YվEHT]8O[y͟x"/osQZ @oeȬD*UBQWz۪Ğ%J/ ϯ^ ?JfAxYbK[{>p=}[ܩg+;ʿW,HWRәL<$=HW7 қ^$U=^[rhaﱃ$ *F+P)4QXO&,zjR@ITBlJѭ@T$m j0(Ν$4#vT \[i/QX1^FIk pQM<~Y!ɎaRg2Lb8lF|Yԓ [YdKwI_!S6]\n=}Hr!%E=jaY%٦,f<[bꪮz43U+v/nƈ Z7 2\LuH!敒D42P+q9VwH;P llk5xb[o;@xMsz14&wZOU2sWl_%CKK;lLj 2r`jVF^.TX7U|#DBࣃ ZծoⓈ;eWG /P8U8J5)8z zc!" 9@a[{+?|`rT-K<ͨ xffL=»O^{"7V UFʑ@hy+S ק GϟGћT[&]rV6f=!,q估nZG3>y8a楑a<no~us3;| ɷLl\*So]Cw)nL|pka/; jϭ+!T PCR} ᓗL(tѫ@.Ȯ>Z|VՒB0 dI΃ ra9o$ gx52iWFk!h̙Drrzq@0Y@FUSmB[Иu11\CK1Ntվ:;rD(>lC M%LXj6)0OH"Z& 4Z+Ink["oN;9; i,4D2[Kpp^JιLT*-pn.Kq gM6"j’`LDKM%&#*"'xr9;JxXqS.Q $ #)(r;w"(5F" BGDDM}+?S; cL6zMt!\H D,+@Z%%qI%h{g)FJG&)+d'cG$csBٲ] A[7ljn#{+>W^nbZ (D*S Ad颱e.W-S֕j:* âA5i4L: JRi#E@lF3]Hܠ;p@Δ& }EL jɭ3]َVӼ]9`~ VK,7k(0=S}:^]-dn4:::tM2TEX]Q Q dRXJg+IʍT>|$z $H9xi ST䲸ԖjgůLtޜ3dj-*Ud_A1q~Fu⅓xe9k:Gӫ_WACZ}]UF{ZZ{gۛ6>Ƈ.+tMF;Tڀ(+dW-`>֛FCW q V;AӏQ7(tB$/$J]dHKHE*'P:LL2s: X_.l4hZyJBX;"D锨#P%9<  6r 3pjskIןOr2VJS,:\7[孼,|᜶ KYgGU L4])֥1 J£Qk5b:ek =S<N{#Q Pᙗ9a^{@l/C7j!SB75pTBm,GjQkj)G[#g<S0mrƽ~y`y =PڛtRsm3Qss&dT @`5ENKBHRE )p=䌳"m \XAzk,$K3Z3&QXI'[#gYo@sԔWGqS>ŴqC_y,X,rI ~$DXkh80͍"z=bnd֮ڡ_JW:oU`tY'7P-W>c"Fif֜I%Δ Q1xΒzֆd ȪwN;v|[!;^bYp室lxm n|} ߀fw'k$N0'+u' ruz`╅bcaw,5/?4L ,jdS'Զasm'-Y6t\٠>N[4:_MWG/׽&:~6t)opE=t:gc9Ľ9)9<;T[|8fO:giy=eb^ܮaf7̫}Lcn;1gTS61U=E6}{FssgmrV<֓7<̴+M]_.G_/y=Ec\hH{CHӌ(@i"zŴ`A Osk{[|~L np5Zho>?HDn-*E@J@߽%Z\ .oi ^6|ULȴmS No5w ]. W.odNG:)x:Mއ)Ls\}(g?"1Hl$Ԟhד ܓp!{+DJemh 4x4ֆQo-s&&{oKĜ0U@ 24QW@s.|ϖط]LP6G?nb=2?աٻ޶W$X}2`v,l0ZeJ#ҷYOE4M4ٔh%]]UY`_~|"⵨ڲ_N_˯Kqϼͣy]B/m7_A{Aۗm ~p~h Y#i"[T«A0 QY+MI{91xIDj@n-MכyY&-s*JHi#Yb 6ZX9ƇZ j7F)*!$|Vtrj;a9[th^sm-?]t-?9 OFE~(}Р":С.h;+rL& t\M9i;nY696[r*cɺ,>%, !E&) =PC'^ѫp`@\rs;)8(D;{Q.ɣ @d}RZSP 2/hI}Zx=۝ƋlԖs<z3xY(ve&|;7( 7i<⨙ÂM(/6(ԝ ]tP,ū>CԡADS!HVp9((Jƀ}ґ!)(H*a!4H6Jk&(WQi  `QI C[cfٍyءH'b I( " j~ >HS;mb zgjc_󞟇9/F?̷Z=ya 4.Ml[yI xQG̛ڿ:x?5kq c!E`TX8eBe[Vau*WZZ+ƣ )Kh%]d eDOafu-lMQ(0Syo!R뒤/dxAưp,OP-3`_/o>fM%U{p񬼵gY߳q7s"I6{6#n|y<9R| zu7!I~[=Nٮ>~u绋[/rP)DonR_=҆7qy)c_=5K\Lq龵۩e)oH7吗xPSېxVY%čSrSd\0%/,tu2*DPeVo`Ar\ ;4D0ԽNo}Dn[?bVVhwe^9]`Q:tvSjm/)*my , _3^HŕAXz_'JxUڮf}ǎ=MlǓ~(}j䌌!@ԑ3EDM&t[㌱` B7*+3-?JURI|*iڟK4Ŝp]]+E㶯?!1SFl\J0uZggo+rֶ6'@ĝ-~mQj w3}_o6|Mr[PvXw󤞸l| Զ Yj>h$#ڤ %Ij{2^X yJ?Սh f,&b!PqFrܞߣ t7v#88VmdZ9k|>Ngk<&3?d$JY0O\& rY @ E6~%d~J_}z=FT_FH| D] LF!G^l=dbHf@q|xvաCm'y#AJ -e,!R"&PkA;G tVD$"T[>4\̕ J t1SHJCoo͇j8 VzNl~؄˛[<&r yBuI ֣ʊ yѡ "tI*WL{crYn8yz(L^`pT+) 8Ț8X[)|Гs1f*rƷ8uzbl@H݅ñ3H2˙x-bے( %ϱ@N;l$БvT;(, Q I8pvE3̛7{NWpMP{#-qXKaF)Y,(ߋT97Gݶˏr ,|,r6Bبs>*e߷ tIm}A[G2(C,͎7 hYxCpP]xh>y ]Ơ2 ?(Ywqlhb)H Ų{rT%g&)ffR7*Df3J5!1leqܷUټ94)hZ<ѺybXEgڸ'vƼ@#Ӆ qMvZAнq7K R 2΢h$\T&Ȍ&QJhπ)4+e^#SQ@HHRRzYtLV`)B,h>`hh6j}Cz"`wi F3qϛ-[n7>4m<\ץڭ6}tHJp+%W-KZP!v d>F.zL-XrVE{#kO/8!ne[J+W< TϵKwr?nɝ j<}fZ]ڤ@5=Wcs LKΰ4v QwEl=MCvLFO^ޱX :qdr~ܣ:m]^n/!\/ӥy+rygy[l2UݦBervGKI_T<չ|єG)5G/Bw:z&v-Eu+چT_ѳͬf3ҭӹw$v'ʝ]>p Hqf?|j@wl6wsXyxWozb/gWw'b ȟ8J.QKRP߈)G2N]d#+}NFp" GJ'yaҐ썸IP@P(sUb;,†TҒG*S4I>9U,0IVTv'E"2{,պC8R+%[lRԑc)BoL z`h1P+#mqJ{ ha |p6f% }&*zBsNZh)q"m%{*o _xfxc5ML aY&R*)c*Œ_. MQΫ 8[c+CIgox 2;DQ;gPVL,Uk$Jl5/6%iےD l&~ >HYL&zTB`TX)exlT%ք͆ۂ7'`ǵfvwk_?;5%Cq ]Ȑ[aUZr=q ן1!ǵt`o#?珣q>wm,Wi%#xe>Y\'/ĨW~OOx{_ޕZԭT¾ .쫺Vv~FA]ReoC%M1J'$Fzg+Y?{ײ8lŻb@v>.nݢݞ([EJ!mE<| %TvepFC|7UϾ-|u ˷K?z?o˳Gs!z?EYOlc">:& JOYy2^J.?7۬v|ԗf@V\݅ ?6}y> y}ȗ:t[D"{.F"mg>sӛ봻͸qYmuʢC%3O?._ex_0[fZ$|BD (9%Δ QvTr~\?Dub"LȅPY2́4!n'1(I;5I2z-~s6sn޾ZN" F} Մ(T< mlROY)\5B)<B7/%̷ύ:]?J(!j0/ѬU޹&Ih@Kxc5T\R\Z뭍YF!l]AkE*bY,=\=-3)2JR P*m\5`Ds&p]#,lHM^Su;#UIO_xJ v)FwgbZڍhD(=~-oROO>/}4v>{>kWUt:DUdFs^ D N퉁D*צYW=PBXJ[ 6S*{5)DxAX #'{BC QZ ARp Iiʃ,e QEE5{Lhnun(sEZEZE9v;(ׯkt**ԓNZo^s )5+l1_G 0o=/Ր|$u @_.e8 TzDP:eB"$FYaHSy*S{eBWW~4rO4XnvӯԌsd;`e`Jreq6ۡ6'Z7L@)EW De4t$l ߳T}3Snv>!^ J3qL ,LN y "`D 'E HZ I8਑!o$7HWFk!hAٜ2E`L(R!sӜ ߗN^l:GJ3mrR!zh_f g>C':aΙK%::FYfqzblOq! & c ,F 'N$-L-ܑkWA"";O2J#mN$$ZRreRh A$5L(ȑ3#4aI0`&&\yϒ48dbDe\/b9g3%ONEG5>,%JHNqGQQF1h(/xfuB$=}aOs _I( :Dq sΐdYz ,RrT˯uTIvBCEcp0O8j2^:D|ΠIZM*"cRE)ɃYgɂɃ/>̀1Cns{>aSr'/ۥ''lr`éŃMv6  |^($EѨ&ᎅF UD3A 5$XxyCRĄBRD IR4z ͝ TpAs x"b9ˉ/I`q-9"=58$F>PBI=>NGP#m ŦsD6_zx_&xGmNm5$Dq@{ ,)a1\ Z #OE߈pYvM051RHB[55s\o$5P@ JJMȢo)HEt1ʀ16N()jJG)!q ĨA| u>)+d/%AQC G/~X#XS$i 3g΍qjqdmMAu hZ<W`4WM cR{SR 퉞c( yJLX _Xc2\ާ2K`Ȯ7ĶoT݁qj&8_A#y4FUf\[H Xc!l\4qJHadԜjp };;-S~4^ J $ ehbIg3h42@*4[$Fo41?Ww O"; i*A4ß7:;EҺ\-:M_3 ,߿RBr=v7AXхؑ4D*OXy)* yW-f o%qֱ$TN>J>BgQ"ٔ|%t`x+S%xA*p$\PR;jPԆ peS8q.`c IJU1$A%%iTQ0 _&<”tS]-4%Jtb-xӢd>X->ﻫBIB,hB@Q4IsNx ?vXǛWǧnlέ#,|zzEz;Dj'g_aq[7^+xc Fc4G#HA]{%= %=BIs?,+餕.ie CڛȥdLXBP DLE jSŬ`9j;RXv|Xl:GF-)ۣh<,g$vw7[M{uۼ$C0C44gnPP97!pZFrύ*JLX6gAE@.&XH.Rg&EKfLұI\b9< xyJCǮ_7 ڿX^WԕV@)qlq"wԺUu$DX<412UoS} f~{`>zrun.6"׫>Il mQ~{_Wݘv#>Xlƍ/~κ%&wx%vglM*n*n&Eʇ tL$He@!h͙D(qeGR9KVPYfh+@v(As[[~l#JZ ghd0 6Fnj8p?&z/g!eǿ+2 rYYWDžUAɠ-1k4 Zf'CzI-Mgq`~d!k[BeLhϸZnJZn^5>/XźO* [1o|4|loq̨7'ga9_9ߝn`4,/Ox),]ؔ{x@#8 ǁ?cJibb<ȅ6渿`wwNYe$JʾNe_41nx g+cQmq8x 9N F.T&}B(GX,@ƓȤ$Ah ?t L WSCZL6Nifo)H&.|ƃwV 9mu L=0ε9[0 |3>rwlPn'+ҍ;%4$p]#,lHM66Pո/U = k>[-Z]z )K_rrs omeZbZR* _ k+o睷~]1mOcrrc0!ҁ4Fi³Z%f7#m;l%7KT q\1Evv??4;t34;5_B奟制ƨdbg(o*D,@'S69 "V1U,k_Ryk=66"چaRܼ5%Ϭ-- T)N \y]ocGrWA>0H0~Fe(R){ {QQ䈏"%3{jQǯ03?/өK1%gp¡h5LA%"Sڰ Fs§;u5,JmIr^ mȫcN{c4_%sXz|Oǯ1MN8x1Z;vf 5?-J͕Ѣs)$`*&fݥ@גlbP@dWֻsQԚ3jXP Fe_ GJBa!յdl:%c( b{Qӛz@kng='4^hp8}x1)I% 9:FUںd!ZNav(@%Q.Wmr-.86 )z#MVH3S&"eL"A..83_c=!C/-X'uvJec@t|J ^8jCu [D9紞if)|hU Qήgf7G$ 'm{'ie lRh.tg[!@jP;@D]b(&t SW'x~p_(Yz|$c&a69`0WtSFnsW Doөՙ5! &p+n2U0i>FMq+m֫Gœ}&__ ٫yc~U3!τ:Q('o?֊=_j.rh,/7>2{ebV5zX}:X- JW0hVҥ )L> 1ʠsN"gV{ 4/O|ĵv jvs-eMɩ5W-> ORӭnRon~9oޮ~6]\p=?1~s߂*s=o8^쯉ʪ۬ Xէߜ.86p//qtr}Ix-]7Ӌal0L(?'k"7|2fВ F`<]by2b!BkyFzM( 6=iaZҺRz./˓ZH^෭4܎Re~~4g8[Fg]2⽕~}۸,loY."P+d_,(nySg,Ԯsvmſw߼ߪEN' d!il ADڐ@D)b2*:;?ڤ!o|AK׹~+\?6<.RQ5 tn@`J# *[Ɉ{Įo:#ص ud^>y??ag\w=/8ICS2PJ@*/\Ѣtz !lL&8'TJd <:z|QkC6S a?f\(2t| tX\DgE`:԰jmc<*NX_[F4mpx\ͬ[`Nlب0F ֲH#Eo&NHq^J5mHJul@{c0RS#peDʠ$$$c19[FzNW.K`V,IHE*ج, d PJIL%+2f\≚XxXZrFV?r.su0>0iRfI|frq>aw7L}ٰݰ׷WqV_J7_Gy` ξߥ&kA'ma˺y1Y U*φ9IW}r,63j^G.u-^kGTrxs13k~uN)Ie/e~1CT_grot򖱖֛3 ONߚ<*[| _`}v&?㷞?՛ӯEz^/>Z$R@'uD8z?)R ur%NΝZt&ۓɔ ?[`%GIeH5:P3h\ӭͷ9)R-QHayߏCf&݇)B(7s|ibΞ"pR)}KsWzy+WOVC =bvz[|fރS=wk_+&Z~a|S˵aegRpKxMN|ZVlHui}%O4_ȷ;Q9l3$hۥ֔kt?9rc= !Gh,Y3(h (Y93xD Gn/m-MӲhybZR.JHi#YS6ZX&ZaØ`R'V>N<\KOh9=+{ߍJ7w5HX}A.5H%"g F(}?_ +oeEp&qin6}7 J$5T֍/,3s͡Bw?^*o4# ДXK|Q ` lvw*-kfu)J DcKIRVyaa-ygRջ+дwW{dveUoLb)}5]&GQ}K!#Y V2vϰ]^{uZ޼>X],B}X|es#)\2OZ4:0hvm1N[;T~NN\[-{lVN~[Sc.C} L3[`a1w}iÖ֕=G[+L b<+kZ&cζF_ S;5<٢kFtntl(/Uw2bM)ML^: :~Y^,V+rU| CfjZɇ Qvl'^7Dlb^L$"`vD iH_ 9.쓖":gCc.ϩv w CB?o1ʇR 3`.Ʌ7̫-}<&_`u%={sw}P%Yjʛ>7 4ml@MI:#'h;' 3䜽EC:#&)54CwVsDқ"T!.dsjw^bM҄ ,JX>Di 7C>fVe1,NX\Z[;|%roWx[UNU_ȀmkeET yE%ե_dI!26sONC+IT& GR,zH!c2Fi_ kY%hX;yP: (TXS mSRXHZYgْώ/w&|P˄H5M VGԎ#!`H$QWEL*g.Y෷BsA>0̠Ruj5+Ա ݶV  5a**HCV{;kBMdM`(|"IVNXLQ,"YS0LJ&HBzQӺR^S 1>B;}=HἷtZzn#s4hٺ"2vANc01g,K*qW2+@F 4"6%Q6k ={tK0s <;\NЛ>JӾT[ QCO }ɁM@:Sc3tXI.8QUniSd}HJg I# I>$qBx!dlAH 1'@l,2 ZnCR(IgJ{I_G0m,߼i hQdG;xxJTnʪET iJF1qLH30 Zd$xeZm9qnЏgʦ>9KUJoRd#(,1BVr/r"Fz+US%4|4!D1;c(19t  ,(͖8eyY_y3ʝfWm0΁B5&LTb\zulן]ǔ*!kR4R""# ,_m 3\0 t+U#j$ ma@e,)HLI:#ƒ(.6fjM-t^y^#?yW'.5q8ga|Ra>n;s:̈́QE[vGa]0m-օ_~jYetzsz dcJ`0ē`%#/lR(R%)$pLQ(jMq𝂬*}g&BXØpXޯ,M~Vhi2l^mf'rQÇ 3|hJ2nO8|ߝfN1;΀ x>_qoY̝4Fl ;|>xW:Qab}lTDp^FoXE1I)lR2S8[N4wdzSZvB%g]QtP5ssWtP@PԸ{p<孔8zy Q QhNY$aNet4,[2 `Fmhٯ0!fω)lBDZRf:ٯV~!m(_JD[aᎌnȳ8hU*5WF_UW*5AdD A࿥*K(XT1'P]>BaGMYRݝFktL(Kt:gkY& qsȜ^nVPfYlX}NO{8Nqބ|B]zϿM7}}vcx31܀os/x^֋:094/ms fŰ\O() ,Ǜg+cƅH/!ĩh7m=;sJpQxM:K>4^KsF4r8FKg2jL ʫMLr"*X}4ѢTr mӉU jkat3 ۏcz6mL>Rgb˪m,HE4FY^mN$RZRrpb d3r2EU6xi\ijEmB#d gZK!~_.^]cif dt?KFB,5(h+t&;ɟe*K8Lw}iҥ…|^Kk9 BtgЀbp.ifLb:9«&-+rTEc~ ήdɍ.Wlw:r)~HEzyD ˬRӍ)7e#) ᨟v - +^W:o@ ]ζCZAZq_~OS/-Ǵ{܈QkKdS125VFȱջhٶ⥤ғ%,(ΔLosvxw~iEh^ ηK뺶Zטt@ zDb eHч8rYJd.s>'MZ[wCSG7фA90>^69G3ʳĵ2#uDGxiLrjbe!XК`NI)姄^DRQ \1e!Ֆ[EjTuW}{ TCR̙ 'ȆFSX`LNZnr' b;SK2Z?JVy`&BHЂ1Inbt"CagCNs PQښRmI %dz05 "yL6gZzLFeݖ"JZ5g:qt{y@-Gqm͟h`pc#XHZVX!)D AC2%)zae1hl8f%A񡃖L8eAA@aFhH 5yVRBUb4а)N:tYΘ6@pH6\&"E )D Qs#HUnmHwiZiԴaGPm3@Z3)]UЕĊc^Ndr[n)] )6tі'+jF:ַ3e}Zrƽ\k8Sފ.()>*6YwȅLB9c7:Xh '^Ϋ'~Ʒ쎩g^xEn]ڴ'jY]s Ǝ/h7%w+芽kn買.Oip~FnBܶyݽgr z&1wdKW2χ|j举cξ^^ڵs&,yuFn]l~Тڹ~]( -FY&L UnXvB)Tls~8gűeE:%zOʏc~lN.μF6K1`=,e{2F0x8G 0;2%US L(N=% 2XW AJ:aIu[c *-gB%d_²e%ؐk;?%#DV=;_[ 9wcy6Nw1^R:j-Je ^kC8\9.={ό΅ӡEA:M|+uD)uh(#+ DK{ 2ց izRR 0vjY, Ta&aYQR>S+>>HѓdIB࠵1DR4$Jgw2NO"a\{IԆgg*!5s1\39L 8\['UΣq"Uz#( $: .! IES:*ahoJ?ֹ\bʴ{6 㯽K:?" 7հ$GEŴiG2fk?gN_fjV"FFΙ0nV{".E9Bi.b+]00_Dz_{m\`p13Ō{-׾R\%uݏk4Œ#$<:oʠb~>~<:H!AI~4_.5ܘ%#S7^*K/7mRȘ'hfr 몈RWLp%h)~<0xO^5cf1#c}Ѓ9P[L\o|ӳ(Ҳ$W$/]!?sYcU^ۖ_@-4 }rF~H,Js¢$RH/:!4qAHuSs%ɚS LKzb||pJ m$e4q\HE!"EwTERb0:4|eO6Q\t)X9tE) hLOB" t%Cc 8hќ C?S%kzP=Q[ֿ`|ӕկWpb^h3̐sR9hP(0r3t]{HUa7@;3K<b%PRh,p(*äj &U|B8aCs=Yz0l}8T>1awlUXI28̜V*IDU<3*S,G)ꔷϠx96kyȎVg *Ul >kvi$(ߤxځd%:+%ajW!gԭMgU, JmbFENYdt+]8.hypR*XV*luT,6?{V]l6 Oݻi18R.ObYd;POU+,WАpcD ݄?,iu#IIԻ7I+? bb*ehKlKlKlFhxĿ//////!KlKlKlT08?B!KlKlKlKlKlKl͋7Xxxxx f=ipSb.ˮnVĬDDmgɍ"9 B'/!4 ^=tFrƝ97VZ n : 8Lb'\UxiSr)01d[ȕdc 9rCPӐ0e>U#[TwenNg6e?"̳/87ﳀUu] fI68{rf)q#ݍ+@hwץO0AC.owmwj?i-Zs`;CjQn)wkz΋Nzh[nCd~:ɇծuhKLCg9>XߵhИyyI<ޘW?៛:}skyE\|sA@]7'mEc@.x,'UGZu>a2YG=:>8;՜j*o./NT1F0 !yI$Y"z:3Q` UI2ٗ!2O1ŠsK+gWPk8bCωkUks< =jԯwT!7H7  ü 39*9<9C4vjna4EhXCJ' )ij6s,DR̵O@*SUNZ!5A+4ckJD+5&ufHI2U΁zv~T|nsq>s%%`9I;*5&2Oĭ(d&A̽j*PS_?cT k^/dWR"+ ,PabRr >0Y" YZV_7?I[s -F!# z8-'e h!,nfPal3U Y储&o$0D,2 XV%N1̤ܵU%&gc1дގd*l;zѳsQj"T3ou {lʈ^9*}%y*sJ<MR5k0#;&!O[Y;{2]&]d+@ZɎ`l"QXvUe+,wҋةY$ȱT o8nH:BRG!d@sҘ-ͽb\I 1*, I ,ct ft-JQ%_UJL@(ER2N68ufz=wgZF,H,gb$z䖤iaRZrYĝL3)'Y7;/k=ir>)H1LڜɎrgT2GfAoкT ߠ 9MC  %-9hh.0hlv.# e yr(WZ똶A!E#hIA䔬9pVX_DFMHnS(y;_n١A[omZTG z)+y}J ֗\a7PZRC)tV\1-iss84YH@2 Ah1- R+P{dDJrdz*Ag4 A &"ВO<ʻՆYm7yſ9v~ITٗ>p}/y-v)džNR㣣ͼϬs+'$V;t@v1yEr)g $z$DҐ^ Y6AgnFc z_ w1Ye|] Fm =1Xlv>x7#ɼ9 5zaUD(EAA '%&mPI%8uygKWq/GЈGL#cgit;/F4ziqq9bLLOnr.j 28^~X-.u7_.ǓW*esRWSa. -|P >Ap%3YMLQѷje/Kq8ApWLnLi4[E!52*'I0XD:3* 7Z"%<AUKY`ʸ^JLo_c4*EoWi2SZS#y4zigOÆ^ui26 &cxs{{{{<|RD_mZ†纎[U [Y *KbT6eyrJI0 ϭZNqD ) |Sm\ ,opX: t7p^TZ/;e">T0˿_?n؃ Q֟ bǓ'c#X)&t'T{|\8!tY d2eM0ћ9d@GHNYAy=g ,%M_ӛ$] f_rMVbsqpuH9D_ß BR%mL/ Ϟ<}"Fs!!w|P'ayZw=݆ЎniYIr1Z=]na{KP~zoҔ/ɩtMɥdtǔW\ȕJ\d#eeP:r֌T˰s_uTW}o`5JQYK{ˣ9 FɁl Nstɩ >k[U̢`h<7C$y-&'Iy+A]lphK%/]sw.qGqGL&ۻdqSEo7drއd˽>&jᅱ\0UdrCE/$$ EV^ y2)sT92d* ^7 Ӏ1fY)SeԐfY *YjhwۗMǚ!CZ'$l,*"dV&PVR'cMӚ5M;j ;Rɦқ^ի8N8k:iQzgTD :MhbBj]=qdV/3ZG/w\jg _СS>ki#zdB+Z̑``Uqų|`IBJdM fPe'ϝ )2E+`v#rh}:`dDwm$G_!κ]"8bqCZYt$%!)S0G%453Uu=l `Cj]o4#.^G8nbquFem'֝l؄wf{m 8WWn$n,6ÄWOW69^}[]ݖ:I>xQ:U v#v F%n/0ȫ<3 gӿqXbBHyFHk2zB+v.sBmH$xþEIxcK ZeRL L8`PlUf\HYֳ[s@:8\_^rVv$h\)H|_²öۿG["T_҇e͠R0I 4}Tğ3mSY.&mAYe JftiΒ֖\&e3*٣,E6e9AFI ͬ z%-/ ktf/(EQ6Z1ve0/P>ЭU6ʫgaq=[s,X`$$%>A;I IDbBBZqYaQy\?N(cC1M8+!ǂW0 JY3nRK*v،R)dۍ}rB Á' zdlb=֪e|:k-j,S: 1v tkr~V6.g}gwn6Rb@0ς`X@|R6J1gWJضTvn)!miǜ׿.|~öee|:ﳋ4VX3<J=naV8Ӡ#+1͢ϼ +j.$u @w&U' {GHH!$}VfhEs {0jde2^0ikچ1PXU1#6)d=UyKIMk!A#k_  @A%=mGs"ㄒCzF`:|]Dȝt\@mَ`O747 07fmj19gphIg$(1a"̠6AhD1|[D 0Og7Rvv)iMWdPJ!༖:9c:-?h0[Wuf!佽|B(~Y tȀ}LjQ9U xpQ\"{]L1N!ؔ2LrUQeN= ,YdL Ad]"MfAʡ$Ԏ^b* BTP3쀈E@XΚsr!'f;O%L&EJ%(:%s;,F$H5|ePO EF3@KZ'٨~PJ(|&مuo[+Y a**H1BvP+-.G?i^1{9"l#D9Fr&{IA@YDr;/68(9-no(O(S:wFH߄!fZ{^ ($l6ѳV))r]ܭ_O=5U m|fe^;8:XDA{oAV7 a1g ^˳|0`u` A-PK -ErĞÚoC )@IIA Wʲ ,T%֎k3rK8+lsϽY=S)cgyiQlY^M` )LTչ$HP(ne.D;ߨ\q&i7C|r;mJ~OO}&!]vggU!>¹}v~hUg8ϓiq2ӗm{xM/׋{|I[*e|yvʏw癭"nɷZ]"߼]ugG*kɥoWÛ7De܌r3HRSYf%jQZY% $J5V%j!7X@3[MkwgrO_-^2USg츩ݖuQ.#=% s #z-rĂ5 + f߰L0maCW SM>Ȥq?>/ mL=]2cV{5Sr[?tI'>^2 ʫ[ jk[~i%Wcs{E+s==l:s͇k^N 3-Z$E-Xmly@U֧ٛk w1^kV|&0/Qex%wB c]hNFu,Z= -gvCYCQq儷QAla7qBDgcF" 5x$(;G(xKɿ919%&4c*=5; auiGB͹濧[0V_j~q_vmO,+v۞&B2lESߖPCb[Ȭc{@KV3 j6a`GwB^w]Kh.% ;o-@̐;I1vgс3B!I|_Pgw1Mx'5cѧuro׬j#je p@{G~{U~&(9[ (G#u)Ĉ]6:OYΧG$'%{.$,zDi kL&SbJ2£t,)$r㕶Hm HwOs@wܳY?YaqP=rɱ;)~ ;)6UEowOFڵO͑׬eoV~-z8ǘB2Z9\9SBUL4 ,hYK6% lKjXPQ@J1Sv 5j_3F)eъ3ՅXQx2}Ưl/Oib釿\ЅWYUgB1I.Ѷb (g&PSAw&_JVUmjW DgW}awOe 䡱nF~yù VqZG="XC`aA4&M@&"1cBf%Bg7el2 pް#Dhy7TA9&0I 12efQFI{II{BsYګF~gˍe9+c҉/8VgswBS+:^u&RվK1cAz5'Pz<= 4E 1E,$H1BƠ֕  A[epRmmCX,n1d&L]Qil `Cj.C9Kf4u>.n֨l~8;c0m7vgE 嚕_kR팞GW|gQ 1_,g[fӥ<VCl*"W/Sz߭&[3$-f73[b"v[%,!qWR%>4? n-)2 3@2Ȣ#yTɜs2ϤB8?}m=.WaF7~w%9KVIYqUŕz=ȣ19^A텠0KnmUUasݠJ07Urߊje^5JzfC"f=eNGJlUz[~\_a6>У5>P˟ ldX&{Jy>Ħ./C,߯ Jئjse>׏_Oԛ?o㮿`g']^v6/ҏ?d٠v^ۇ`7koa8ԛdS~ݿzf[Or >&:Ľm.Zls7V'?[؉ul+H ħV|ts77 |q=itXZ?{׶FdEا]`KO{yX e1ڄ%R#ݟߓEERr%FCX̊ʌ8'+OF% }9^ 2qwuV$%UVش+!KFL.9X&.Y{V RBvX%(xs"$v@$2S 2)4\BnxH v+ (*T :Y-X shlguF( BYRTʃl]y7[SRtXY;C7a*?%QvdԏI L tfS.% ic @f̤C@8NVj T|MEw&#%R92rܤ`5pG 5Vkx5{VP1bc 2Kէ ]D:-d ՜FdYl$DO;4չ~Wk?Z6i4֙p U_ZEb( 9i`#BTP]?G;L')A_0aFYi: a٩w7\V%mɘGޅHМ6D-$ >:FjpdPYd:Prt9|htVsc (Hv9Mkj,(t_XK(x$2H&rZ!d^1@& .LKc4o=GDC%YZ{[{#f@܆`m]ů VfESY_ ~ygD%lLe5Dw <}XV1l5`j/L1JH؝4z 6B zxX@¬aFX&]@-L1CA+4S"p#&D98A=zMPO,6T*`$I!*MkLCicY}6FJR`eSFvL!j^7"0ts't\&!K#ָ܋#\R!wTu)Xo&rۭ^}#mu7Vw{J t)7s&ȶ_~}C?"fVJG6ڈnTV~Om.r5)5R%(`/^ +b%Ϡz8iJ?10C~+3ї \hb@bGma#@"r|6Əq6^G}1s|rѓ3 Wvd˧gN:Hs \5׬f5kYs͚k\5׬f5kYs͚k\5׬f5kYs͚k\5׬f5kYs͚k\5׬f5kYs͚k\5׬f5k\u@Is {=k\Kf\JXsV_ b`%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J VUiuYYX:!9lR^4d !iH#LRU9*&Koh6} }Ԟ˖9}v'?ŰBE^zuU%aѩ eb( 9NFu(%`OF$eXKmW{k=9*)b=܋{)llUˮn kTO]wDF_/L6~4=w)%>Ҽ%lk+77R7Rn^߹)r'=(w[|c;k'˻7ܾ\ Ҽ"y7ؙW#6WS* |w+c(6WH؜,6g9Ylbs؜,6g9Ylbs؜,6g9Ylbs؜,6g9Ylbs؜,6g9Ylbs؜,6g9Yl޽&%&mQ\^@HT!J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%ϣzk=fѭN?_}ړw6o7;Z֛}gcot"_p ;zK07Ы. 2x4K?p[土s|vcHb>@ d<4)0 SSml₾?aA_}_|AN-:g֊$^% h%]T勗GzKBt6yc}g3xA&6|헫T]!-ːW_UpdZߌV~B _equB6izOozuy=^g'yд*ݺ2vy'~3nBOqpW7\4p}E:ܛcr[OoP~ߒZsaZwt7h'1׼yK7~K<7k˥稈nWDF^.sjs|ӹycfG}[rL RPC&] >6ﵫLX/>վcEt_>/Gwg}]k4eA=_)|i4d-@GNvbX?#|mМK16[U41꒢2Pq> g1/Ij(*rزs=acPpuY7olI=d=)}Z$Zp~FM*$ZrJvJ" U'dٲ?,3+54^ycRSч +!BJRn}oj& Z{bu'Ҽib7~t m%M[ܤ0pU UEIME6*& ZNؖ6\퉛-06>Sƌlb{$_nVX rܬ?,_|u_Z,}el]۰^|XŮ[f]^}^mgM۸R nnuY ,gqfEua҂J[eכfG\;{0f`q\ޟ$-7-apUk1}^;>6}g_ǝ;ի}mb/^\o?%~y_}g[̨owZ\%C[Y*gJ{9_!i2Nva{ a`B#OI+T7xbY2а$&x"*N(uE}9f?ٴ|y-l1!&ŵt(*{d|d%䃈g~8̷l8K,\5lĵT>2ѯLdåp'\~$%rTl3 |Ck w(>0m:zCBFxS1gxe Y(" 멽Pg0"#w9uw=!`H1+VWtlvDp'hw9=o+8x^*CUz2=V-gnpPqLRQdapNZ@Q@)Yp{Bd󖹽 4e6U}|> /kIY {^2D&kBLMH=^@cJq6jckin<̣=tq ~3mQZ '3vV]l؛ۍ5J7I;͡ux 845TtSΗz!4d"C,ӶH 40nTj!l5?'MwA{ =}\ה(ےEa4eQ%Pfe$ L,reϲ3%(Mj"4lI(G*@(wu\h(jtSʓ 80]Y 4ʧo>o͔'$it4 ׹S8Z 4VLbJ6$at%~ fCC?Bu=ϠWqA/OOx*%Jt/_響'9#zOϜ9M*0;l}gxs}wM5(|0jpߨ?~Tɒ^Uddd *p~)y0~|t/^w ;DWr<Y-x & ^`Jԍ+BJz+SK\JL oӉkHT"KV^ ְEm% B5J *gܫZHNH[v2 t45pF j4Bk ؊W8\+p.pvoS\Y$l䤢 }/N_q؟K6{]sh |JGȕdW QtH)ǘw'?OYcFu'[&' Zx%}A!մdll8%z&1d* mòдYeeaܠcz8%6 ϹJ`Xo`"Q{oP%)J6V^9We -gLO$66Qˠe2%SPc.]NlmXb76; iN<$.McWڮ-]=ߤ4*`UdW AE6(kRfQmH3%Nd>Tg}lR?Y&x4)|(9kV"Nk%7-V#QmJ pEH^HYg Fd/x.d&+^=iI;v''Ċy'jӵMjݸ#Y-zo&{ZrJ쫠bH.0/3u;;r1{TLw]C9G2rYȐ@J'ҥD-qs X} xsb8Y89}^z CH`*yR!cV{SQ)Ұl|LynY泐}+ н@*VIb!GWV`v={}N`ILW߈J8X&ۄŷdv4_;xboJ/v/ozs9~o^8u{y/Lds"B6;{<%l)XNI`Ë~ގ={q,{u5Sw<M|jr(0a;?rYT;~wB_|xMFc_Vmk9l2? w( Ps,lQ25]4J[JFJs.P)QjyiӥUϫeY-Ȳî%] +~Ot_-z#e SJQҝL/9_8ZرQBt[%-W S8v"=*G+uЌ[ԁ[Q)II*xU(xCQ[(-[9 i*dMhcK&їDt+.papelI{p!XQ*8W$"RE"pXx!1L=66n!(gDɋaڛcXѕ>t4ӡt 9V#~X1f8em + \ِ"F.,WY$[z&lKJsZLy˃6\}Td!Z2ҭĕ3y&J:X7΋<~ IJHїwnl\'5-}> rІ=Uw;b=vN}ϡ5Kz6.`e^rVزv31%RDyo@e&LJdqY57;eV510P+c{%!q \`=!T)pp%9d&Ɔac2<ם^_HcWk")V۩!Ge .Wl˳^..UJGm{BtDp=3P !l2!%aPÐ ,TltGX~SFgn%JccAN0x"ǃPtԛ!^zk@Qp5UjgAw\ռa^?== t.ZC` eL-NDMO۫鵮Y( kU)dj[bBfs9I2"!!tBX|6=H]2,U!IypNLPdRɀ.! r.̕:W0N-,r&S~oS] UC2g]dYwWe]O#_8` _&.v;?OmYǏo0 }S&g9w6g4k1󸛳H_}A{|² S TS[Ih Y?{DYẗ-.P&-٠.3X;3|F$~w>wqG?L#nWK)D?[V2u%t{Qۅp?*Sp`ܲL^,BCQEe;4Z6> Gxݞ垫&Mo5C%G.33S&2cRɳp'\~$%r.. /fn"_`xNxEQovׁx77{gsWfϸE96_^C/?kNݝo'= /wX/bTW6ԣpDfGL>=ox½ hw9=o+8x^*Cϴ>[gb%md.037GeBe$ sբ+. '$/IIN +GҪ5 ż[g[(,$U!KtٳLv25JHd1%ira &;X _Jp<08~^)ml8iE]Wi-Tn >_4}mg!^.[1!rmVI 1ixZ eh F} gwkknXuܧUyHlW*UV$Nj" r,.A,@U)`w/_tަ@54^%)B[=AO*s"dIf7nPQ[MR DS9LELpoȋ88ТqW>UC\$MhK0p.g1."14ćϊ>OzO4)o,Nm6NĠهHq:!Yic`0+Fv7>3MΔo|3nDVŔ0 F!ZJP!H')#)&FP5& 9:%ԣ֑>S X "^0I]98=` B]݊IJ$SH=w0z$z"@'kARmB 3@KR\k[1))I%G=K1jǸ4pmR(2`2 ~a()dֈhܰm-p08^o^r񕔷}:2 *c*!\[-d2jN]É(QI< $p=FT$(43 ӉiY}E!m `z]GK|ডxZ~\}Q)^햷IcN0.~P\/R+οTɮ%AV`rlƕE/>G@~(-FU WW9XER>+QxSB3KqL D^ K-5{d. FTEN9-m l`T|.(5,#)(r1܉(@IL4Rd&D'%ELQ'd`1^i ln`H.Ԭх93YViaKRArPO:?^0bYD#>4 ^:pDZiFu!D)"j\WR,#6f'=z^ I H׮?X#RCِP79d%(DI'|*YUtەσ~p@JL]%3T̍ rnt'xI7]( =S9(9#餕.irϼ zIihea/B&Bm`*6U \E (OڴI>#erFï)hG_~6pp ==&7Yyn{z* Ehv]4 XMY d]˄^I,[(*tTDOPoI.R3HRt, UZg9 -]|r<gl)j7r_y', 0`u93\ʁL-s?p4;p+F?p>|TK3Q|{IUX y(.F^ϚƣޗCD3igǪO請m+ny-f>_@]4| 7!5os͉+u_cGڭ_,rr0`e=Q-lA9~Ubq#zup^[f~z)hl\ ̒.䦓5ܥ)N@VӐkU~#Zo߭-6ޏ jiCKe;\#`vy:\; m V'/,[ayaj`֫ɦtQ;Ұ@d!k[Be Z4o<rVB['D ^WG'Ux?φ1gi::3Ž9vv|w6o1{9l"$?3aSf 4[ncB;1jlX*.=ϺMjz H^HS݆@LQTJxb*'ymfiF'\8~@wWl"s郕Z) *p&ȝ "D%@Bƀ)6A'r,x ]\2 !zM !֐{mZwW7-tm~| 0VM˟snZX,@-\o?gߦ^O/O^6QkzbW?tqܻE8Y YᤞI9OZ2]?Goxo=y*O=*_0׶f=&8{7g/f+3Om~]670?B+m/]ڄwY]x&[-u R<860TLWu>D1xƮbG7?2p2Rܡ}GVUD䆚+"Rӥoh A`b h&5!hrrAe1.7CPFa-6#mCv=;%I✣ Fm6τU3 s!Ƴ Kв̓jhٔvֶO|w*qJJ[2F,4MvbI+*b#7-Y-(1 _ڋ ZR\d$TiD0aeb7k(sM!dt3vB1/jW^O-JG9[0yFkchävS'{`ʏ|f|wnfczj)}u^tlzH|Ư}eYzmNMw(⢁AWގ?LSC g-zZdG trBCiy_0R.>&_|D,0πTg@9@ \Wɯ€/%з̓zᙅ<흍l[Q[pp[`k9hEu!^Č:7tmRk@4&Najww/b!3G=G+- А!õd{mk{)37Cm'Os8PT>.{_=(9Gȵ(>sʼn'R(R n[Mzv ǯ9a J ĀLT(?ĀL%]bL ҫ ZBq8}?/̘vΧw SwuČ+׏s;w[U%wi̵\9"O^ :bb&dZDTƊ0o4*L`49E*ZiKPG Rӟڍ\/EgjTۡNhV=mTË/銧iS^p?&-VE.PZZ@'0DB-ME-3S^*Y+v`>{ʖ6[%3\l[eWJ۱Jst0 ӍqTLG"Oi񭋯D[:աhخ]kq]zzCZ'AZRҁRHkCQV-eAY ]\@-t#{zgU>~֟yDePv0h vC "h&o}NNu6F%18A :gDB0:QY-gS%nRX?>|ڽy#EAgV-- T)NhBpT{VE{EAa(G E2>4e(ʍf\\[9|y.6'2T#4q}m=>zd>zL &o^zU2E˷E:ޞwA-vȮTYcirZ4Uh d|Cf6#k K 2Ҩ,֥ R hٮ t !9GrCN!ǂRVǠR)lB%UZ3#gf܌RN qƾ B£EkcwS(QX`byÜ@*}9N29/n;n.8dDR)^柛xK#>ɼ+gO;6l_ tzj5-D jeb䲋ѣ5ז-Sk.و#M0rE+6M|v[e|\\Ua`=9ktyQ6CڃqǾZ4j>j ~+J9K$S*JzE7:TZcN,Pl؋)ӬCvEG1jbTq`}89ިgb<8V#.5pSՈ@ͣlU6=Jٱo8d2j&z2ֈFrS5ы7:v%0Q/pobw\o9Q̛8 f{,r#mȞKv $.8#vC^)T90x\EŔoe)51wd0q5˩PR0y==Spxp\؃ C=W>vSV!6ͺZsET8'_X]g.ʁͰ.~QSЎH(§FopKxXv;̏n/Ek59( Jx  g} h[ zVŠѤ@-85Ʋ+-ySj!U>.7|~(F!d*o~Q|dKP~15;xeYn?滞rcLZwML,'@,!Yb[̉UP ;H"7uϭ@4-ji_^ƕcMՌ>hW@ DpA`'vHG GgbYg̿NKXk .1骄9 Q%M):SNIe23ThؐW@&_g.5ִ iC`B:t D\ j,S61RdO +H9Fl+&pcttE\5'e]Mll99DOW~s9B˔?E(?L6kYs1/Ok7;xќ]7^w.K/v5\^n2O%wMh"I(O/v&l6N;x1Mkˏ;`hhA?eZ0oWobäk3ªj=c|6- k~n^j a.U5[]zS̟ڠ~fW1kO~][[ͤi8Ds,u뎗LTW'ry>KZc!Ubt0 ?eߜ{4(0J]&ٌl.O[^[}N,fJR"4b'e-Eߖ[|z1rQ칿qo~c&Ǟdm6GUߖg )e[YOs&(4h6 (#5P lb^*E="6VȦC4 \VV )$ۦ(gdYDEFkeS[c4i' '5ٱ'̷W}}͇Oqs̵5Cx{,1U*B*Z{?\{ ȂEÐ\PAdtx}n'?V*:4ʦ*ZlMUղ9&Fg bL%&XIQw.=WUB4AیF;u*O`~]#Bz`49$щחtߘWr? *i}[嗋~o^+dKw_:*v?a_.eWQḂP0XDŢs`[9P#Sl Qx%zdwO;ԔkT? |;7eR]!8~3wڼ޻Zsw޻ )E"v;(60؜. KQٮDjQs{^%ұcHlŀ gc(;mdUU"V%mp8o|Y+ԇuWn~ #!CrmoQNFFwO/bȤ1f?d|>zf/;̼1a~|`Շoyߩ-\N%"1| y*+yR֬z]S sw{KʝT-6%)pG4&erf1'"y.Q(&]:bISc9{; lV7ӕ*XJ ,BjʔY%f0bnFU1\'VMgGn֡;2LC"Hk ~H759;xzHa<[_*o=ȑ}Bz) &vP_?Q(B-]oթvJR|pIREg"G!hs.4J` >˳OG9;.bbhw,`#Frus5 U0j4^EMyl5S+rնJ502pdQ/C:RVAJ+N$n'2 2 jd!Z Bω"0|Ywӯ 5ViTńX- +"zUGP#jtF;NC-]-!f.) rP FM!)_m$ Qq 8t`<񓔮6N=Y,/GVYYvQ\xl1A\0&6rHQچmYK " XVǰXo))Nd"B"[kqc8XnKsW8d_KNϳEab2뇬~rp=h'1䕂MUMn9 Iq>zU;t鐙D:pk4d 2Xxr? aWTR J̑'T`W*NJ&,2hT6$%J%-7^') ZF:De!v>ހƌ ^U l*Pm }6[W(Ek4W!hSYa*Z$D&G-?T(d5BB< FBI?WX?s=%6{n)Hd}y}2k۬ us# =D!Qx1HKF,SlKd%U_ObaMxՈjCD{m} g^P5}kP}^ Z4ɽ:B|+-tBQiBqz6KMvv9b,>o 6!֌8f+c:fE]NuF)ƛjSBeKg=hl~:|N&X']\ 3؀/ui<;gc[w.(WLAlz"ӵjJ@R! ^ݭ6n[ѭ 0uu)OnTOZVh/&spWQ_KuCmKX]|z%}s=M^w=k)ڴe=֟j^ٸ۝[|5/^Z`2gz_;mxH@`7W[kYRL|oպXZ[.UŪGCXz!a*[z};ez&˔ R6?֩akH& )ZFG!m%Ŷb[RKݸ̇DZƘ, ‰\ U4pO&zudBًBL}H<&څhQTq*|$DzM˹>heGb8rIGGQĝv8~z`KL2fu}?L,xCWDSJx"RBCN4S@FeL ІS,ׂy@H5-o@$52|#VG#$%NтR Bb"%95`R %Fār x d9XO, %w*.Q;d֜g^(aPk\nx ­\Jn%sd sɲn%t\ݞYZxcؘD˯75,ĶmD66wva=lhmL׍VW)1[$[rHJnۆ.[/8]{RS) 8iS<㌖'nxhsج_nuӍt2Au\&frխΰMMxsTJ9a]V]g_6nQ4g~w{@I9;:EnA&,w,.̗Π]{/gnCvpg08f0lsPq)ϾJup!$sϋhK{뇯%{~\ ˗{)!sBM>/qG WIL"cYR2l rZYh]u9[d\hg$F ^xPnQh2|"R`Ē +b7_%KMVFLXV:[r,ڊ)\dl&Hsva"WݑFYP|> 6?2ﻸ?|o>zJ%N7%'_{<* hAjy4)(.eShA `r "7B)|T hC L* D6_T}^n+J-@*u]pIf.G,F6Rɛ O%/.d"JyIRyg2%щjdN%ͥMyM) 41Fojd j\]2q[.RVA}A9ܬpJî5 y} WCC5)h[nJ4#Ie7|L.*1M2ٞ2WPL5qw2)a̮PfwyPeYڻf̈́b? ]'UKq寯뭭rk-D9e54.?nx@윷#3*)2! ]yohS-4(-֥S H5RBrR2=dq4tJܪ$Hf AidlȸYK9 6bn U'/:+9~.5%&V%:w[~+OAa{^#,"v$:cVFziL$8"HcRXm cl9GͰɤ;Ƀ&&Qn"72Wan`\^ j7iMڭ㕠9F4'D cc!q@H͸TxT;-!p= &D%T{#ba]H"FuT6+~Zq4 bSq "R,""?6!I75)qIR Ӕq52JP,jT\*>2lsSlGĥ,\lKE0.B-.^4dW}tyɻRS[KWqz%'UDQ^mhyB+Q 3Bc.)VX !h*4q|m߅dRmK]D $:4N0 [/"AD,Pa2VIy' O;39{]Rƭ Kz[3k6z޽9I+sqڬm:$u l]`ju1i{~YfD6f!,'{獷w^_T١煖a4lox1;Ϻ-br%.QWtӦ;66oi8|zhs d,6?mNsTuT,e1Ir$1gh9  mE*H2TQX\{!/؎=ٱ"~zl}y>;N03_a.%8Wuǝ*lLq;KYe)opǝ&?3b'n;9E okdu Wwе;sw*dz?rZ$?g)\=Y5O,e߀Fc/; Л!W1`Ɩe|rrM왢l[K3"GƁDgcN 'J 3r Nd^%bJt ?㧁 (D̆(:~;}ȱwwx+3ooTw}[?DPF˰r.?V('l+n}{8:"D3CeIse,6rpţ'E}M휕qϻϝDgcW3QCR7*mT`.57GyLD6f d !42L`_T^YjnomM5^飘M20!Q (jp0".E LVJA{T36=e?/ZiS}L(mD>AgJSEB޺}[~!V@NH,ECH aBXй }BB1䱍lNY[ JlXM,&P@i*KzĹb]5H?ml5+ e3_9$'e9=' <߷.8vw :E|8g9C}.FLP8LOLrn2矇b2  =7$gDkR}Bc'ȾVߪ3m`Sh D 4d'H@tOK@xZ~k *#tώt5=PRΧT'S&T,̥w|wqa CBB-.bJ`dyfΖ7/X`gbrgYE'FRނ#s;X6tRiH&[7 /h?;+W% ;{աk|`+ Qwd8DB6Q $9}%%,4wZtx<9a &H0LeϹ z %`Cgɴ.pOQM;D8v$ꩍś~ z&[87kӗ9#9lu3 HbuB+`Q20Ho\eJRKM'Bž흂bf53'q}Mʊ+&EEjNj6ŪsV PLٰmCt66L \J ñWI›\DwqӾa)y9r$1Šd(G ךzA5(cә6nEDe{2]_r_oyʷ)߸ ɉIT5םb`kSEۉMHzzSs0vާuc yv_"w6:dZ 9pDO=$$&I6 7 {f_{L r|PJJS4m DxʜyDZSR7<(_if"DFoYiMs6bE{N~B/d֔q xckjG]1BO7hȷ`$K WݶzE ELBXj^ermh,*R555Š1e;Sվ65|:`i-܂**\㢾Gk"1;V !Kª5fultDJ甝.tݵ5hAr=Pxhf;Zw dgqS(V&ņn"DMs e\qw9l[.>]|:H9> ȧ.:u^~Foz/i>=ˣBJߎUUwg%{nr'%~R=J`JֿeYJVN8U*;g߯df=С_D`z: ]}{v˟6Iz/l}Z}X]~|sp͋ڜ]|XϜ~<9Zߟ Y$;Z45{ {b|O /OEprIb?ɛd 9nLJY([5=;4=0ӛ.P2M0I, )c9tQh.zLN?(TSw9։q>#*TPGD2I؋cdBG%3W:ݚW66g[b۶HRf}.L;X^, _ʩdYt?r8qV^eX\>^- ;:|%$s{og8}d5oIϐ~?~a1!ОM[Ŗwsd)ާV5u5iz4}pWĀOĒ'01ND'ㅂdSؔвpܻGq(&zOz/&D+fL[,5 ڜLJ%ȶX*'+ܱ8.ݧت571վ갻].]}Sxv]T{| y{ñ7iS%#ԪNTVNj  }hmcX[8p* cIeQq$C5:I30qnkq'ǧmV^.W5S9Qd] nsm`朝 fM Lsm`60m`6kd9&WL@LoNۀ@ @8M,Pn3yL.d-\>iVȣd ЀYEHԐػXf+k C$IEHR(:>z0zA`zSfd)K KHl*\qV>熌; sؖFF+ \)V"C#i)9[ff^(F+Bˁ{, a`kڇ )DoARaֺ ݧ-?堩[V.l=bo6o%]/ԝll`#4.ZwH!\[֨u8VRR*Ϛ;Lsohkgk`Ŗ $ֈmJJ\bo0! ճdZyXb bv!#QOm|(5aG"j Boz6L;J'5㇩نI-l04g$vfAP@QiVN|l JRf~6w;+P 1^jg6JBž흂bf53'q3fo>>+Z!9)rٔ`@aL 6llP8!c͉MClv|7#gIS JHrpAk"Q( 438`m܊Dgȑ";`20;=vp.`GEmOwۧnt}>mOwۧ%sey%]+ؑ `FGdGZX!ɬLNv#)oR?F8'^6.m?;|xdsjwM_K|4m¡vY,2Z2­'͏9>p,},Drer#QAqkRY3%!K, 9G Q̕)w)= dcJ`0ē`%pt+cPRh]&s3~TIB =< KJ?=(O.Kwjۻ)ھyDtY`C?mxd8*KiYI0 wZb2 񉐈[W [ι+;:3>z6BO<2Q``F3ދ.yrB$T^%! Jۜz䩂<8*rzsU58axZo#"A^VYy>I$U'*ⳗ$imk8%>ؠġ]z)m R2Od%Tċ/=oNG}0qk"/I ैt \܃!k=G.+cDʗF-KqWmo^ i mpyL'w@$2łܯwj6nwyt)[?-yHq<,l h[a3jSE_;v]0[n$v]4~mEmD̿H]iKƑdS1LbF7hpG}ʷr -p.Z X@wa?u?sn“ΛU^c#jU`C,2Cp,%g2'F}ITХrؓ[gG-fr+M~Wic4ml"w'b d9 Q%W%#Orjb bA` t"*[I+^~KE!!*h6V6=@\~3%XޞT5jrrf9lz{=nũ׷2sZo@wSz6ۥho8VגǤ**> ״8}|) a.;41:!J^J9AN1@}r%5\`* "yL6gZzL U g72ng)/Sd2  %BrwmdOAGѰ}MP08g뭑Lbh{RJ&n$> &CRؔ .L. pN爙Ec+#vF05dՒCQUFmףv`JQ„ E$cZ9/LL69ϑxf`fkgJUB&fȁ($dQ 6j"Q)"h|RՆ;~Pd|ؙjxh,Noe}ѐ-frIF֚̒޺$xO\I%UEDg9Q.ke2 rB57mQzFzS꓅pv#¢\&%⢨OWuyؑbCݝWzfMyݧ}/trX}^[ZGϻw}~eK68iD}4M %:#}G:N@=bjQECd(!q(9ǒ>K&L}.y ܛi.vEI_eUt[t: Jޅ[="(v[!gĕ`v;{O`*$/W~4calWJ0~l4b z+Y1& $2+YjQDkWh ~#sYq\ 7 !Z# +1{e@~%L3d_5 6A 2~ g@hq]F؊ytr!kzT/C, P*K5''j8e`\f24nO) }NQG}S \ [>KYVϲMAztF+#]mϴ^iqδҸﰒ.NF4I [}VAX(жơ &yr^`)Y gkU kRx(L䒗IF.iLC\ YR*XVq.PQwژE 2vjٽJCSq}F-W۔L_{7Ikl훤$w;GO1_b 9]_U|NimH3獶C~.۾5Z_} ,6!\Ztu|_E f^ٖ:gY'25sjWa ]NP͖XzJ{gǖsI<| [5ff)u(!g-c0`5O9Jy,YL(N=% 42XW>[o ue=$\Nik gOJOpm_N,#Ɔn#P@3C_.헞ԘZoh]uWgV)25!I޽gFPA Ȣy[( |M&nڒRMwR(Fa9@4Xց izRR 0w!@hEFhwzo:NKYK+>>HѓdIB࠵1T!e"Ɲvk-M{) ώ,_+NS:5C$B!d8)r"P#(Q{M7I8qu\BD9t(U>P(wJ@f7g[zX SU*nm;k8ng8M;gם/8}z6:Xt1~/%8ѴLLw~+JbQesF ݜ';<`F &a.^O:_yDz<`Wy|٬o{$Qep-ޓ\usWS3N{yu >~, >?^4\%bTܺ@Fodu# .iGYQ4g_J1v~x|zKv&C2xPD-=d#q5;4ZJ.o[ ۳vF\i5##GaX9)pKŧp-M0_ Jx W/H|'E ϋ<_nZt%bFmݍekWfτΊ- =\Kf\>'w8Śݢ+M[v QCɚM4җKʐ6Jp'sE5?;Q5b?y֙8Xxۻ2?{'\o%m G=gnBPN:, INIn Y!gz[}} >- ^GH3ؒlY|;jvUX"4fK#bP#r{ihl^}?v:\#YSl-"f~>ini^ŗh5'B^uҡ__I5O7;zkgMt[<10IN5!*ua{gYSoJyo.hB z̙C~z /EMgo ϧOֺ 'Au/{vc<)b (i~$RbY*w}XfGaQA]rBl#4t5fr刏m(f`ۆ_g/ݓ?3ů`G'K"y>|X_zgG[Ưjmˍq³q򍽱+QKjӻ<Ȗn!o[lF&;/*92]*tV 7Bcs6ߓZ,]w־D  ePEgiP5P# h t9DLnIg)g㨈,s&3^fʀV EDC,*ȐZW7~m|6winYiJ\%J ҹ xuWQ.O8*ƃ12eJB"MuQ:Fm+g5.þ7AB{n g1sp B(ku<*q:=pLry Ǽp zi6ډ\ԇ}>/\ٝ.rg2W0Y|SEtZT9Xg/+h'y7wG+Ł 0qR;,CkahhhQׯ~n5Mn:&ܭ>L6 rwxv#ى?=/&?lZ$c >**~}[1i:{,fZZ-Rm{r4/b=PTc*t5Jz6A0 mם "v1DՉRWm)MIR7%mJ:.쓖LAE:g]pȉ/C+C&'q"ۥ޶BCxb[ƻAS}g{G/^q⮝ ȝG@6 =s6#\\nj+,.ϝ`:CE3L\(m(m1J:lr"HĄD2{M$ C#Y1JEuA.v#&&%VMȶLA)ۇhuZB뺠V֪ RUoR"i]T6;Qz4V_*wdEQUH G(=RVDU0Q\"{],)H)㨹4cT3 _z>;dJP/KA& 2j @k_$XMᵱM,S-I$P(Txڦ "O$-zLʿy%wLT><0))A(y#h0j$AE QWq1Pd4 T2Wv`6**e|y~JR9v\`BI0RE)FjO-DF#D9F(I2'> E$+uJ &ɍJ9JXWvCMeZH=/({ $A]+B;S;F{#&c kB]De|]P_b0GS6V5 CPX L R6VuW9)bYiE02JmڌN=U!hGTpý |ɼ#!W_Y$j:!aL69Pa G7JI\pt:""FN%))bbƔQQB C <!dtBɟlsBmeB#A6%%9/Ӯ&c`xP; RTKD gXPCdcLq|vv q* bj-Rxa2ɠKHEϠu[7@{2jp4 i,|46i\)lG LOnŞo|˘u٤ m}&X$} 2GrE IDbAw%Z\Ca BDhR>Y bYX &^{(B(e֌XW*`  H*ۈk>pX% lhXB^ KHq,!,%{#K[d :`R|( 6/ @Ɣ1H>tx  C.b £-+L*fd ]&Ĺu0_!w{ޭxaͮK,'xqs9{ȽșxXh zX~-1d!2I?ُ Pb0dGI@dTtGwz:Z9אj˴_V;'ֿvg/vo6aoi8I(w)iZ/,cJ:]St%hҮ\iR]pOhr||{+iRS3&_Oh:qukB>{nGrwtu#H6_ и~O*I vI7rHvkdxws֖ՔXN{ m"#;CD۞.sH'Osϳiw~>K>9 Ew)vF8y(4b=$*_}3Q&*%*F FJVaa:YM:3DKyiJ c4^\Pfh[)Z=@1T1S`Pe [SQ5Ӓ&C(ۄ6"Rw>mH1B!秗+!f­㍷=ad9fKN΁җUk3WA.:ioddmrB %`#m'|ItN)`XJOsߊ^g+yFh;>EhWaƉ**βDM|>Ā%(Q@.)BZ8c.#NlTk6 5ؠxgحgPk!f4͚Sū5sIPE(q@ɺFnP-h͍J~vg~6ąUCwWxm9J@\l.Hi8G|WҧnE[1&b~~ւ9)Ʈ(R"wt{P0yN\i>qLp nF;=mkַm׳m,78v\JGHu H@*ˏtQh5ֱ3m!zܡثqԫֽ9:z݆5e=ewǴAJ͏ tm.#:rR#jO+!=]WciOW5n骔c{7ӞN;g}3>~ y99Y|>N{ 2n_9ԲϝNˢ]!eHF2JiZm_;kY#>1a~6\ #ߩop# D*^"Qn1#JcH"rFW.Y^uQG{X|-ycFJ^~J.5%;.$ZzM$f>"K1%׶,p(-+EDPA2d"YoҠ7F9S>Ĺ_Em#GE] m^fo6`g&RL|ql$Ϝgq}ˎe)rVH)6]U|zލ򏼄Ҹ&Zqeڗd5'kgU{)ɬSX&'Ov1,D{M׿fA c I+i% 3 RBYtl:Ēu& ZJ%FQfZj@7.2ʾe VCa#5el8-(Yl`-Blᣧ.#kMbfH\jqY.vz2~MPdQP@PTth ieJ*Z{=Ir(*Fd5ʺdu:;v"5F[b6;r_`ұƁ66t1*YZk0rѮ8\$Rf獛ig~hnb )z=dFًNۚL"YI'mb.&N50tN"ƽCǎq mb0'T&Ym oE 5m*v+(3x@e+s^/W:s^x r^~_Np>?M3K[1 >^InחTR5G?J6bǪ_*+`<y 0- $5"yV|Pg.KmASvXM,}~~Y 8}V?^:8n_}N%8x?~C&RܻƫyOQqKV|?d-Vvn-+ߺVPr)((m%dz6l_#MwY{ K03[k,vr%3Z;MA"`'{[.8 )YB WJ 2^ "+ZE5/O!It ^c"}9pLZxu'ޕ:3?#/[}a-zs͉wEU_Zn.y]vU~yR!K0 l6HFH"ot! |E.^Nj-ʍaz>=Q6',I*I3t:G7.wFN-Opv񿃳LHD(I5O*:+V-?usW?=sJ:avgڮPͨ%-=v,bc<9ͳ-ް\(;뙘ИvP[1Q5?<ϳ?ʹxZ/5W8Z~krvM䎭 G$b?n>4]R8+]~s[DWwօJ qY@0xzu8"N_[V̖$|pnX7_DsQz[;e|2+Q西Q/}=\* tR%}sx~>ϕWmrvsus'8+{=.X`+~dѽi,O?)UQgж&G.^2G~ǣ=[^g>h~غ?:"/h{;<1o:V$[Re(fQ`"IIY=3xFdCnO-M׫9_ fF2u1˦1B  Vv3&#%ċmӼN^L&vɗc̗o'"];XkmЭBwx-by"xXRK9,bfpE%CgAH#4?Ɋ5)d=($O6|laW9I*dSզ"ikP#Astn6q|GGUɸ^Tq]ٞկ|y:ͫ hFkHن,x(o*oɢ*RR LSAw޾ݝPkVlj`ńnͿy&FKe/Б&ݺ|Z}Xm/$jWՋr0$V[a.yQۋ])֑gsQPƀW ``!Hau`1>8>Jʢ]qE-XM"Aާ"ctW֡Ffa -> LNF^*D˼"ܿb}n{|4L\.h>rFz^8;ml.`R"0X b4JK!>{H> 15ap6z=lGai_Y8ﳂ2*}Q7^ޡM*)O7] Oy7:uu7k}DSo Cg+ 5ntkI\s}1<;Er{X|=XVDY0@(b.Q,h RKB0)el;vj'iipӳ B6d :AP: 2*)`?JA,JZQB{]wlEKxʦ$1ֳ_R*>,A cJ"E"#ɐfH @ԸB;d(uS!UׯEz ;_0k>Y<^_B.DQ PN %d&,YdѠpP+MO[Y-f!&10r>GGO&(Hu\15&51+.d+XV6˴*&%h)H GZ5M اV5ip;"<"TEOE]|cmPSX L RVuW*yZ@TJé-A;] [g|ĸI7YƇ}J='lL&Nz_P==0b/^}WG[=o8˧t|>Prҩ(s+X "z:GtE4-E,d89Q4bN4 D6.9hI8d {~\ʕ^{a tF52{(Kحl8IӺUy U-`^Қ$PEIѥW}U? MSGSd6&Εv'Aժdm')./S `3OȫD7ҊiwG"QjIQI1IbUTeqABF&;-PjhqVRH,I+ 2I-:43>Gt 4]2xݩ( $-,:d9 L:Em"бOƫά)+"ɼɪ,Ɂ˯OOz׿H=A ~z$h=/|`u^ٸPs҃fjxxxۋw*P-nvw$I)_|bK]_z s1:kr@8!*hsIq89V7!Ge꟰^k: vCu19U42@}t6d2m;:Y ?t` %=MG-VyP5j;KLFޗǵ^L#n?Ni-}yxs;Ǟ?VSofL:Z5WY푦Mse-4%jD)w}8g5JMIZy)'oCbp1=e(sJFUh]z8-Iv)}w־F/ X)8Fe 98\FtϞƫslz`N y~9Amh9av+m:6ns+\Zx`dTqjY{ʖ9i7]^Mj%*n"/liKHrVSi]~<40,1OVQ1,`r"#O+MARJ < O־oo_^ w<MF_BX=]kB" kNT igQdlt 8urMͩ8n8~/Ll ON9nz(Į{|cs#:[<^U\suZ-Nqء;ul5jX0ӻ&*ݲcXno-z0o52UZ."*o2)CZ)aХkQH#1ߞZlu$uk}:?:{NGbXk%ĹYKClBWQ;-MQFzEՎNGR]>gkj)% ()Zd>Yp928H}"ͤCH;AȾ F.DAxɩ!@t]d yc g<د Z9}=ˇ$?Is WXr-| _o^ZG/;}*vd{y/d0-yؽeS2]P:`هqF\XxrU\œ'W48xrГS\lઊ \JkO~JW4fgWU`TgWU\W, ։UR!\ rV+0u]_\ӿOKlR}'6O;Ym\Ec+#3GD$FUB/r6߳2у Q>aEHw}W{?t]Vz-bU*RnۏRxEЩYRޔNR^d#%蓱'|ɪK ika:ƻh__~W_yN!162ì)xkeYHCRΖJ@ YD/˶LIktquZK?T>})*grr]q(A$ұEdφ6XOB\j^+3ov{=c.>"X(ѳg-=QkU!Hڸ  YtE05Դ!)jBc0SSbbQN$D&$B1FC>; !Sf0hjS7I:ImRRCL/%?{WƒcwiXE``y>Xa[ߗՒlّdnۊR&Y,aRkb kfefR=>*/.G,N=OA8笥rOvOvZ/ӿ}7mX/v5\c_Lߊ}^$2aBi>xЪSQzˋirota8_eZ(Wϒ a?1}>=/k%~n4 -˯}+_%m@ jgv3/?7y#hbQ_ͧf[dfRe/0)|2W?ESSnmVUڌH{6YUJ} ZmQ";e8s~W1稙Y)N[*Y&U7or#OKUy hǛXĈM >_~Gϫ5Ç]JLǑu'KmKT{s"~wcR(zMQW6A"Y"/aQ*VF%bKFx)TfDn/MכxK/VsFi FlH$Y6UcG<_h̲o9fitJ'Xld,Qqo>lBmh.mFVhh?PT!fUl|(_\[Ȃ+8:uʆTʶ*Z5y ((y GS`*1ʠ0I';WUti+m1WX!XU!ip, a$>ǶIݪl?/~mq΋ɖ. xiZۙ9^?UEdEӹXL*Ϟl56Fx-</O@miG> lD vo^) iLg$-|bXo4(q>(t%BF)Q35( ޠAI8% 8D]>5 >`@%Q)H]F 6nb B Di_]A6* s5)kDd%u`88cD}Hh mmcgȬ,iu ]v'Eo[#|%=Mf>$"WJSɢ7N|tպ `O^CxGxy2zjU "r ꌑXVib ]^?PsIN[-1%ΉUrkd#^&ΞCFoXa¿l~S/W=;Eƃ}Hw=u gG"͵aAE';_-F^nt7c8yv~BnO~7'_1p}ycջ `oyjoi^}`bqr}SIWF5WoMyy\1>}6O_'{m6q,uP_פ#;{mֻ< J)@tf1'"y]",K:GH@NF:@)7GU\)=B5eʬr-.Zņ@NU0x+QUaW %eM2%DJ1#V;Ї_t49g+ikgdp;MR-?tZ>pTՁuĀ>Tjq!Bf<{{*LAµlp :$SޑX"/hcdcל^Q}>l@[7 6't+gr_@^D4dmՎ"b眂.1Ʈ;Vjl uWQtB Hքl iF(T(|KfX-J8KUJꨫMC+,d+B"a<&Lgadu!ZIz-3ӯg4YE#fqTER׋r47j uq+#ߘ^̺͛(| '@ L3sQ.d:U[_R 4$DJ(`-o Tb9$*y ]PҰG8/y}/tHb[/iw3̫=hv`XGM6/#y%>lG9B}Xx[P7uɹPO!XڹKLΤ8!cJK~[p3ht~2<:|{b}^żr* j%x9M lEQbs4[t\)R#ӫe+)6l*)\q“[;YKOZ| O'sʯ&&>]m~l9=^~n)o$tOE|bEpeEX|Ykҝ<gTſGppn/ިfBUu۶c@=/E!.WEe&uNۤj}t^K.b>ۼLp!i\oDц M`-"Q)FgPCt6[tB-Ǵ2i S$$_C19bX`C0D1oYo[sn潟־ƾͻBm)c"aV~Zcl>@\2*YH4+aeU>>`#io @ۛ7✋)IcBm)ɍR+@b58tDX}v)qstH+8?M[P/,*VvGWg}D:}qR \.^.wSZ sf X~E֩uVZGԶ}W#4]>:cxW4Q@ AMՆ)6EE$HSٗmINSQ|Uր5K:$A.\55:gl-#2$OhpM{f%jm_㢦_\KרGOwa@*v9o\ !DVJύgsy=?)zF;(-xo9_-Wce:^HCݥ= ҕ,~n->Hyzښq8ܠ娇ɘ.glଟgj{ȑ_;HVC;va1c$9b[-;ȴ݊;A⤻fW=3W|%kt тrW~+-LJ;mE>*4*fct%̵Tdx)SB SZ8@!'6ታak*k6ZkA 32NOvBXgքD֜*$lɒ).DAFp&GA U 帛PZ7Ӎ]Z ~tG~6䅃Ǫ!gx3Ljbח=c> s2OV2s8oOr|-ڨ"tY'!tEeՁ#gʛL \R>O3peهrOLe,W];zlKf*}:\57zʽtT/Dt$^R]y{"[ iClG>θ}IQrQU)ۊtF5moPt!]Φ{F֝ȴ^i1‘l_O.~t>-??egm%J:,DcuWZHЪ)3;ҾJȑڨTEΦԉ8ZuE`ba¨OmQ6ݭu9oPR5_}$TEDH]%xҞٟuAƻٲ&ǽ=^G^GXEDRIN&Х!gHZ-Q8m7gt1ڦe e (7P󿍬J٫(cF6<iI9uWVcR |_ߗAMQ(eOBX񵲊few1 jCIP Yɶg]I jJfOhzrĈƥBA%d^J0HFf,isdށ6%4 [o?Q=c=!(*lj뢅`!9E 2-1nJ;\)PPviDޫ(V~*k!!N\qR",go1/g,l]| h$ EGcM"ZFđ1I":;ߺAM3q~RJ8Dl;"2c"7"kADVQ'ŜZ8vH Y둒%*Kjύ-X'urJřBD6[tRzN{W8#⧓̯:0.VLKvE刋#.ާ2zaqTY "$ &`d@|gEԈOCfڱ+xw'*,>EŢApc E?@G?9an"I#T!.u&x'wwaua;$KxȊ:K4X{-+֗Z՘ 9 +Pv$n{ 2ː%fEMx0L_ZޒQF. t+qwcr֞69\?}9T'_`>zs3`ǎ-+^oE:U'›Rɿm*iu wT@X]~-;I%o~#^S>_6Tȳ,tymB _[>w+~dž%<šoEˏg~V7Xb݀?3{7ɢ(' _;Ho&WkQ/%]v~(#b U \Uiy=PrWW$X#bѪ+u}c_ XJFw[zF([VoWzY&i5oP 7jW/qj zoE!JGzptWk!ؿOnl-UwΏ&ֶmΖ[S0#'k_0^&3~9ϘNeHg`ڧw~n +!eEP| "tI*WPfMFmfimYhgɑ\L6h|(EXH)Dc֦HcujhSZY9"=6FU sf0!kI7flig]cHfF((*~[ڮ"% c9X$k;MfY"dJD .`4&{%Po/PZub$5p)[S(Wj$_?0REy)FՎj8OI;fZb2b)xf)(9:b4 YDY+PĔMcWiJ'v)X14,x}3#ϼ*Q1,0ӵ">ÿp629@ڭlW?>vөj|jTnb 02 *u{WFi6d%٤5<ؠz:6Ӑ]5AfձF$1_tC{]YO$jU@XfK}CCgW{4~^!m9Gyfb5%R1QD^gDE٣6mt6DmPY*%o t@cК)+s VHD"Lme-e8B9XatQ)eJKXx[OAbW&|5  B71TzӕƢ6f %e]V#d|! ]PmH[: ƌ&bP,\Ⱦ'GUrvAi2im+xBdFMs=jGz[lзmյ{Z$#MKYAa]q;cjvdґGBF dؗk[/E'pˍ'k|Jjn|TYg 4X V.*dF(%[ޣў 9S@{<BBˢCv_53YIM:@3qh>uԅ ZЧ}f\履"_ϴ.2Uc'ͭӇҬFzr#;zv zzvtz93<:ssҀ%ge'#8 MHQIcBb*  `j fNSf!>dr<=g mzw7WyK 5It">yurՕ;m^n}i8]];{MUAzuIOk^L\n9l[[xs JM*.})[ubr8vzrvLOxެ9mJtXy%qx^'^ #V[ǝER|\vn{vk Jx4>뜅JqHcNd@yG# %k9.p9nD8 "^>:0|S):l^琥h]Ib 0I/Hdd4)sd jVeibJbf'Ji*k$R .pHsȠ8ՖGdHa@%Bɵ,%M>ۣ첐UFl'S}j`MMo7)GW !paxRo#e# "@ц9mx$*dDbyۖӫwJB cbP2kQնai`yJR /|ǰvk;n*d E*I/,:뜴fAYR򔂤qZ0Ӳ˩ۋ/KsPOYzB]C]}; j!>}z9x|9}dӆX̮p.jyy!f "~Io s+{;k 29&ڊGHJn"^$K&FEiD)MnGt7u,N߀e~JJOG/ᄑkx?747=N.{eEK;y\t;u6mw|jƐn""j-A;,%IAAHL`(%c )*ii*)U8x :!y֬N\d-ˎP@YU1_6-.n_~ռx7} QuW SlT_'zݾ6|aȽW$/dyR2ho5č1+M}+<Ĕ6m݌eTz ܐd 75/%yDHH9M)X\,Ֆ[Pˢh椖ü￾;j|rh=b~<%Xo{z F;;J]Gb |_-E몢(`8QKa1gIQ J8$GLAIL+Gl B'=LbVy UR瘕r:9hPVAݲ I3b1 &CRؔFQ l&ێ93UFjٍa6MLSAjq(jʨ4حo!HQbecF{Al9ϑxaj]v9kc!" 4CȊ-aMB@!ő0>N,QjEx'@D\fUCC3DVK#|l1+d{O dd,O,ȵ 9h_$nVmZI\&D.8RNAíDi `K3UeD݈ҫ):qɡ(*"pqm#0uBd頙DDAG rL^ Lo1&11pTw2O_#wz_1+(ο/H&e鹟*!V78w+}X/pd!߿)7Ř>w^QJ_\c?O:qծ!^J#vqArѭ0I+v՝uA.E_e }%eE #s8p~c?cn@shavu1ZT}ČzjЮ&XQ|{A~?"c7s(w4{#*$qnv3_/z>%\?_VJڝK%J"4ϻIp1b"Fj!@b>Rx?^z >h-óc \69.K&lv@.l 'Oq%랔z>qW"<$ܡ7!uf΀(%dmZ>}:\ -W&yl>z`*MFh_ VsċwʼS6vcg>Tb ׆nTWWWVMoq8,MT258йa S5۬Mw tx*]YzA9LC|cu');HKVU*gckdJXHdzQ*FX^9fK63!J턶"%h!B|qJ$[CY}Vm9{&CT{|=UY86ғ_m _]}yN9ps^U sox-( mk{e6XPG8R\zz,dz*dxޕC* M&NnڒPi)MjNXmQWdo:˹%@ A@Cv8IUfwXܧcPE)lZ~ia݇1~`91;JL>!ڢJH#D-^F3ansmM{ hlן945H]{LJps)=^y)TWjJ;] $I P$Tgt$e@10mr@%Lmt˙L0QJ]Amߎ--.Fn$7rZ]qC5Y=b1]nD~FS짚 fn4zwA\yk4=X inJa^, uMJu`wQSĪlM1܌'iъd?/nA]b4Нfc9v./2hjkb7) <':odu34hG}qv/bK9Z?^ǭaH>el Eu{g\2qv.xmWarw.h:ÉoV?ad1rRCȬ[ vAKVNSpv2!,pG ٻ+[aRgY+NwB7kbRrܐѼ+IOOSEl{Tx=_]nyW{ʴǟ.Fϼ=:-V?R*m/)Ώz7>!J̢$)kɝbB$N$X1hnZi6M'>ռ$d$ǔ\ dm)ж5Hz@6>h7֠ET<zO9Z58${ȎqO2O;kc<œ.XCPŗ E9ZoG/Oo ACq" R@c8~CxC&zQ~&Wjv9. 9ĥ9EQ`22+[ ^Uz&g't_P*_={>wTl+?_]ٻrWyeO7oEV5O  Av)BednI,=%x*~EV}%adiA0Fd}RZ((pP)tQg%4EyW-"~\?flrߵ1}g9;(NAmo8Zߤ|XO/kl8S!M.yrhgi'Eot'N" B1h.k1 gFAQ2X<IAiGR k"JP Ÿ@*GM>'Q+i)JD$!b,*,<81"[>%ȶz3NK/j}KN#9H ><4 _ ̆Tq^ OSRm?FkPŗ1 ,;aH9ÐHY dK Co6oao+-A uR>BVPF;L5ɑw6Y+ƺZ/KY}6.1 UyYܺ+_]^?b.Hǭ;nݏwϯs6MӰmrˎ[SϷwwrVyp}yoǼϚ߫|\uݿ= QoIMYᯛ1p[6/yvxbskG9JDp+:JD@ԫ9J:z-G}_qGUJQ"Yi|E n>`^Һ7WUJl+3գl'㚫Ik8qR:-sa\3WOzɟid__^^M/~/G?}t -ִo5kO٭j>M~p{}9?/b{"Ol,Y&חeă긩p$Tʹ#gRXʏzN*Zuع;wcn܍s7vعع"5vع;wcnܯ{5rFUX5ٙhEyĸӸo}[V5 |49(U͟&ӫ+m岬}^<7N\_ l43QP PW .fmm1{ZGEtZO#JWc|IQ5f 3;dvo&31׬:Eȁ"I|h; TNZVh;ZVh;ZV8B}sT>Dc w0u&Bl=1$hDʇ'Ym#|؍N& DKI&9Km_4FYQEq]pѢs\s#If Qg  AJ9 I>οKntPq2-y0cF 8~|)>-%ƿE5JYT!/=:AT.IEL1+ =Ęih1g#+2ec1Q),`DE"M!hH:{ijcG3/NieV:Q0NP?FU̙ Y Ozd=M=hvy[*}*T2R(RRT"Nɐ ,rY򂨡q9PT` _=dTP`!O/_ITV.V !1D0KhPk<Oj~=Se xF19LADɛX0ED+uJ EL518)vaxmZ=/J}Q@^`pikhp62cNXCbӶNDXʋZCx秢|-*YװX d*y+1^ 7ɜKIkc<c7qdjY;{<:^qorc{ QN -H& Q+XV~|+NN:3uu &HZHh I1'bZH($ۂ$B@ |<!$rBɟlR$m2;˙x-!)QJ}(v*H#(002(@F~AIɑqhƩWD+[;/С(2{&VnI.7&C$:ߋT9_dz\jw^g4t4rĦ`lҹR؎J]@[/hbϷ A2by_yvqBF,Ef҅eNǛK\_UYvEf7džՇGGӫE]>3,(9& >NZCj9_ KC"|1H(F͞%@:z,"_tHGV a[:%H`!mx7y5}k#7}},ڃA? ÐsА)}mC-ti4>]OMj>Mkpa>żrM4~P;ixoll5ԻHg- t2Ɠ̚Iu 脆1zbCh(.NȼAϛE|ȓ|?ϳɂՍ; )&/}o~97>.7|H{&: Y=k&8֯źۆO;s8??T1ݝG!ͼ^Y˥׭^;|zGzu2 ʾU\c_KA*^PJ * AlWT߷g>__]>FBRu:3D$ъ-b˭ͩ=FzY!ZGNjtCi C0(.b$ :Yk4FbrCʞcG([ؚZ卪I&Yb6 zh({|x)]]<2;VB<~%[ConFL8_({ac(je%xjD)wYAAQ)J:IB Ųʰ_ŽsI=~FYkb?RDQf#Qu4Eؐj=FGS?ҝrG_6/Pf1>8s`):zɣP `8`фCtJkHuioo6}-FQTEd GϚKb eHB^uv/&Dn6 #[W. E9I![|ҙU/^O%oVB-F>yZzzޅ:/8T0KG=i- 7B0UW6⭛E&ՊJrUT&iR.QBIKoq a&~.0YO>fj6Q}@ur׺=n>{O(/g5_{ AR7Lzn}z_V'ySzx;GȪ)qjVFvDj|hJᮒ*l6b.].vܚ:@ۆzt=ڡ{CZuߡнU)0 ^naRpyEV?}/t_g2/ی_"On;p#]/97UUgZrM=?Vv/ qz{|LpwQ c˃~}^;AlxU1vc՛[{pezqƳOZ[bZ&*fd 2J2>x*^i Jp^Hr*Y Δ7RiOO+FN3ҝ>p:.~}b b=>fͩ* f}EN` 9Z|3zc9yoƝXif&O;_z&a:b=TN+miGJ#Vԟkm#G_E_"/awowp ࣘ"?F3Ywba!#Ӷt$.U*փ8MWmGJE< DL&XhH1vEPyKɠK CzR9yKtώYgz &$?YJê$/6r[6(^SךpݫJHB A(ud/SѠEtJa!Zܡ$٪pՁ:#P4׀^n]7@Wi<2еlBHX7?]s-}ǼҢPde蜱C h(Óoޟ;Ҿ83I."]!j{r8Px+*钪w!d,]ZKxZYFŜB d9{[@ U`glim3rX ^O,{>J6~aޣ块KQR [Dc8aIA|LiSLɶX&KM;r)YC)S aMbl!sVx6#gF>0,I Zr_B,ɮ!vOB,NgnQ^}0O0x|qȭH}tn_-EJQтaMc 4 nS T"J2A`%c $ mΌ%=@6e]),eFj-]2*da3UƲV <7/ˤN/i9şЅtKlAAJEAP\pV E&,*]RAt:;-eSD Y TMel2 f+}m;| )H5.i^31Ej7]o, {/[-mJV a*\ [lmhAZ].ؐӀl!32d+:),k2d$I GQA,dCh.{~"THVD\d5f k^*!d(aY{@Jbw>K(Y_h}Svȶ1iTEY9&dȀUS$ZlmF6.H#>:qɮrQ60A.0Y['eQBG-P Q'" @o1f9ŧ"qǮP5j7y"lKEy^GncW;w7D?PU a(=.y5rVoG9fIA!!B/];tq3_"e_=`MˆhQSČ-vZ^0gv19΅?`};:Qa7{{]V!qF̍˳Ɋg˓j˛;UJ=etx&aJRߏ-!LWVvȌElbQE!<³*w{׫OxNN1 lV6(!mqhB쵎l, IyW39\A'7T:3lm g㵬%Tim"!}z4ƛ/M̵\e/_=0&#W Gyb^O\>~c:g~m?# >5{1du=~SǷW8yJ](dgPI:W9^s(qpz'f?\G?G q*QDv?߰{!fso=yGF|E61dBER.Ҡ$օ HG+3FiQӵV?¾A삫%<jn(_t%g8>Olv" k#o^s IMg쬯3 vYv,vVdiDJVl ItArH`Mt4idkM oxޣ~j-얡ZC;ӽ"^]4+ue:m/M J5RﷳlRw7l2WVk>}{k䫛[4Zv09;gM|Pе|9Z'6 5_z;])]{98x^$ 9"Wkp&vkݨjJ7n|xWT@vs~cDQ.Bɝ txi)&iٱdn7VUmWd+[wn@-RRTrB`A@i@6X@ )iC(Ѷg01)H)B&DM.&z$}ĕEnaьzWt_Wre_kʯRr:V)ly ~=yz"\1\Y N&kTs+@w56yҘ"l*懯:ХŴp&-g #er֐V:i1b Q',A7h7 ~~F})Qi>mVқ!!"()0Dc4FH^iK@E[^%34 TӃ6d,si5#5)3\{(SGcѣWia#H;%NfE~q$L0CJIQhS")$rwJ*-XzzrZyBQ:(;n_}h7)K7=pZ}W^?~{kXհUF^(]~zōhGMa5!ShQ>,UG)Nck?ׁw-&ciWJȃsq@0I_LgNhŵYr#$@"IXGw[6͗W)A}'yzߝ̛(_yEG.[أx6VN~ӂM~<5pRIMr~Ju3U]>s}gQGk;I/'O 1Ӹku6;rpϣ -k{+VfXw4w?>a=FC~6 S]m[GnnF(fT$ R+ [E1+kIJ*U3cQ tum=oB1Y" dmL9Z(a&~!WX08wt4Oivq7[.| \S]tuʵ1Ƥk"OZgGR< FE;'A !-D^!i1פNaۊ]o『^t>=::>UvaR1Tl7tNuDbt$"=D!{6Hd}Yf[X$UΪda6>jc@}`BB+D?xxnJ}D;oҽ̙Txƚz|f_K|!t AasæGJnT)TgErQc)Γ ^dbCM;ۿSfPn1פ'WbDʐ"A/Z`t(tm^ m2ZDƦ kE>C!Vz!b0VZʴ"g;X2t\I/Dž6e<>"Yx:WWoJdg@6BqT_eDslu(HA(Rms>;A#"a;K.fp عw_/d+$mEc uf=|z IT&@@+bȨ1&c4E)o^mjgAb C-*PyZh*uO``#iP7f/#y@Ç! "%V9 %(Z%vD FT$Iu PI"#&>p **>Y<^_BDQŴPmk$+P 3RE)FՎj'~-IC" 9Q$ɓ'}(HV:tL> pI1BH)u4I=/(=IA]+B; pd}:aI[U /zq(lk|m7vaNc3 VuW2 ½"6%Q6k y3+2vU:\cpqb1_iWS/@ܫ'5d0˘y:[myoM$.&KHU\>蕫vܵD$Pkr(A]h^mWZ$ۘqɤR>ˆW0CI"F)ffR K0*f9*82+<u%}1h\|@{YrJmP_'oΠ3 u~/Z۫?),~((jl~{xn>J|?zlٯ=חivXՉܳzIY?!eo~X$za?kݸ$1.F =6m>b-4>|X)Ǯ܄h~|X}}8Y5Ӌ9:9wb?&Ū,eykMӓ*;6{;8U@qy>Y3z*!l4)e *D%PM:veU֝/hF}W*9Av1+TZL,'7::ikOC'KL>g^*& !:vvy 3a*ՅDDYKjMF)PQgkAʁ)@EK3C + 9IOQ^n֝1:~OJ$g>NO݅"V1?:6vwu> y8embJQ[~*Aʘm9X(/vDGA:,yW Fzg(VĢ]"&:ȼ S"9\ڦҼzWӘ3V%z+`&wH涷cqw3Y=ߧ׳-`+"QYIS E\ȌD#{*K YX0O͗yA! 0q!*Z0UI*O$3*1IN=(X TiGȟns9ջ,lʐ}0dXcȚԙ4H~y %,h"{cI-lvWт8T2t_NN6K-@YD^el)cReAan'!7 6,*伻DBz z WT%;1hdʑJ 8&!ڞ.wy'fx**ż#LdqU8DqRYcVS&KsY3jݼۥy/vj/G5 'sR(o^wԻYպޥu0{nwi}xt;`[tE]ԃ!wz%l~X>i}x!%5? uC5uP Zn]_'7-do2 bX`onIvjnjNo۳E;Ԣ7=!P_jP$s:Jt@ʾ({ḨAw}lVWaqXY^t\]LOG6Zu_n־u>m͛<ޝnԶKizRE~6Eh}* [b5Lo m&y]¸DNX5WE]Y?TZN=NeR,xu C X)!\H ^`Mˋi%`[ay[L~Č֜*^E4Q=eh3 3a&WG`*GAz7HJ$ztWh UF+Q(U?f O#x}0-fqm>Vڝ7$0OAz)ّAԾ(b1=yKz٤{u97up`퓢zo&>io鱳nTZ#kSu^ج#f6] 蝾z4}ͭu?iUiQ>ҤJ&_i:YyusB0Ln~r}SЫ_~ξLXmmAzw%]n|`s^_'W딯ۺ"8zq WCC\ Fwkv)3HIoNB:fK]i_) yt6vӣO 8^YStb¹{-EPR sp4 F(BIY6h$S1hť켄1!$q+ >m7mOF{g)wbyec\6&sZ̗Uk3'هsFrI{0RZ))YkjyP} h 2A)_(HW!tUIufݹ߂`lYD=[z_ڠ9S~Y>g=lYPp{tX/&.VSCTBv]Щх|$܅Bx.dI$B4Tqf4+l(5Θ(K^TEMM h E,O04 dZTZN=6@hTХ1m֝]sz\~9˳ 1v?7#!q5x+{:#mAc4&Ym\'1#_}"т^I8#>IwW[eǔ>u>V}r~~ZYcW1"w t{ !O$ yY% i7ҥ<^!hLvHBYt$J1[ R6#UI[De \tɵizSHRH+R]Ѳ=j(EtZA@aȥi&ȣ^T́tg@F5]Hn3#]'ۑi=2λg|Wj*{ѩT~oks')2rYt:ہC4F]l՗[ o*R L8O @&[Q+tX'[Tn:@dM%jL:bD @R3"u2PfEeo6ΎѻwB!?z}\|Z"$咝̠K)IZzMH|D&K1%V@8Tm-`b6)$Ad#qJv@hrtNPu~C^3-yU %~Ѐf"ɾ"@tEeMJ_ݚ<&`<6y$bRy ?. I3q)$U. 6nz?<|ժ4W?G޻,L*b%1sDяUoa!N74ufWu6ױۯ_x|8X3G)z?/ 7:/agwÌep1u׺|?z#0]:K, {Innp/؝>L.dgWlIv[jR,Ilu]d>U,V-Cl$%2qfdIS n9}[w0X͡ АJ>Mra󅙮|χ ]8{@+U נZT@n!1ռ6r,_j5at2w@jtJLWѽeAϗW/+շ\ԓpyR-ƿ7Vr-wH`!pgઈ; pU5bHipȕ!*lP".]+rpU{~ ;z|S<9\O`pu/qxbc~*4pu?)q Wu_=g]+k*I+v*RJs+Р.|++{Q\7{RAfekVp8]lZ@@տɪ?yQMƳF^;fXq50R(& j).v3AI``g`aW`H [*m0l`Zm g\q*ҚgEJT{zp\ڝ2c+ +pU} HvW@H{kMEhku&2-b#AɈwzIJTlT_Df6S?<Ġ$#Xj`(jR֘iɭz^ygq/qsٮtmY+D~,pU]479|K+}y9!WYh yNQLB TXc 5 ~7dмsPm~Oam9$/W;Ecm^<Xo4/޺7M9i(ǹi֫/t. `BbSɮDhRLQZaB s܁I6aBh%ƃ]wWL^x8h}*`XdbOi{^TÓZSY.?\d~3qӣ<=Xs8VSe7ۨ ئP/p^7wJ]WZ"0XϹߦi]N4c ^AMN6igYT[)=XVfJoq"[T7SwͩkcDmU&>XZCeJ\4itZ9*^(R^1<]Im$e>D0"UFrT!M%ad@ll,2$yK&"0Դ}J+_i*U_\M%іC̙.GY*[EN)Ʌ,|I̋l]=1{C,M9B{Mskb5RཞjlJ +թ8K<+cJRRJRD%ٱRh8MJrDa&Z!ȘI2gez֛8ѫnS-.?.uba}2`k9vkkL,e8M X$op{ث@*pPw'F^AzZfb҆L(SIA$ˁU.f,g{ 㽝ΕwxbQ9b!E9F͍60&^cf'[侯9NҧYǤOmDffY^ˀ\bR)!s-3i DȦcVamyDHIjz[Sa)鞕o[ 5"qA^39Lp\q*ZR:J {O=Vբmg0 z:- kp閺^fT<ۥy"Q.oKL=D1%WOlr%^4BR4OkE!%[{m53U}HJo [:e>'ea<,qs+0 >!JHd,'KtI I ,H#=4VY)!2 gZʥ(x<7q:<>'aEWNhP:d,--̒3-L Y R{"Fr(U6A*RB([іKIvLBA4s,_7q:NA=bSR1[tܘ G9&Rx92,rdf V!!kZW"JKmctƓEtH)1p.DsD 4ٹlK!L8o՝ ]RWĨim)TY+>kR4Rl=)BZΣ$ަ@\S֪mMR()t᷵wƶgD@;xBq8ߙE\aw%DVmO\o^ Mw _'4n͸B.꛷g:{WgYAm]lܳF|3?/O (ҜER@Ya)<6F]R@#,^{$"灁hK`7w6%GDFE PJ[b(ulU8Jq!rAZ"J6'Җ5*>KBo00Eϔ7qDWQh(њ]f۷p)#XPgOĂn^牨.r0_1 - ר<@l}?ؐ P9^'l핷uJ›{6 $O> p~X6>ZR׳%E`*Fkd:,O D:3WeDk%EC챕cDNpin:s||Vb|*RAף4ft_Y(ҚSseU]6 r}ݫm~nޏ>_G-牊쿷;ɮϸŹkPv? ga[ЅkmBI08F_,E-b+Sz%a>Uua[unuXTYLNpDZw.}ꓓ8^`ʲF| ̘QؚɄ:y, {g.<Þ2[tFtFk3gOAcI)i#|T^9 2 RG}AFZD )K>&G|X BgBS-r$qL嫹6v&N@ىw>M@(ܯN&ó׻tVM#tyG3ڝΎ,ּeI;0^4F!/&0GR1Zg8pF1T2HŹLL ;ؒ18Ip >q1qcP:*wĹ[oY߁OYn[?|/ZbS4-v޲|R .-[3$jWዴEJQե^ݦn.ng2\-۠+=#T1x7!ǃ %P%)E 0c|ػ,J#X3R?f~yK,rYK +1[WR"5ңBrJwO% Ev]x@ot81#@iRZWz6%&:TZb v7S]"Z/pKFg-%Sd;ʒeFZVN#{í) ςq-SS%/)쁧OfGq=гG ySq^0@ʊI[jq RA: ,(W#ȳ&qWbw@d9v^MdD60NDTJ3T{LDE\v8Y?<-Jֳyk&o2hٮP]B݋m ҍe!dO|/ qf߽JiC !BHs#Bg -&I@[:LHcr\ C2s4" RG9pQ nJĉj>JQ/ipj2jƠ:rx<G}FMԿMl<`jq[%I2>sѥyzMuja $}-Qd;鎏,.cpi ;: 羷 ^F4!F2h?q$ɟM明nm Οn8Κ9Z^2]y#>l^!vͲW#FU앧d9.fK.-yg*t'kS G,v-uj=ޱٸzt[3A]#I.>aV{I3lۼ]Mtt-#{AsKoNY+8Q5w37|Or޻Y)G!x sNeIY,C(Wz}i929(3wd$aWWR`!)4sFhc8g=A^@r"9Ч G_1n1X쒽-^]7p?&EXA3홀(J9Ч?ϔ,W< xEaI>UWZ -/q x % 8Z˾N~ugPVke(HL"d3T// %t; $Rq!UY)xZe/G}V5iq׋/Tp\KYVcvuGF&tLʔ7^]݌y=hc*H:g :T4:T JV@xhi }`Z:X&N AOJ7DKeڒEʊ8 9ݡRk 'geb .VhcyĘsquPHY"!(O?#gd[!XODWpPB )%Ϭ0JXjVP ,2{U6gG #F:*q$74ςO!Z橏p(s@)E/Q{&6;NsSq('XqϦk^]ȓ Ըs(*uJ<2>g!3‘ٕL iu)@ u8ZsD! 'zP!ǂAV R6!HFbFrJ1,,b.£-xf7Y2^lB!ACokߕ&3Gl ˘T;ФVk#V'-,+6El9>ǬwqaK%8 A1TBݎhF&JFbFl7TPwڦ0j{ v++seYk D c!JҐ(ɣ*o @tIHeӂj(R@-kB$^ȴ(8"Fօ$TGeK2.Fn9{Ϯq'Go)qY!q%t(IVyBHIɌQBD:\{!OX;ㆰ aQDku )* ZH{MH)FPtZҠ2FK5ր&*/*e< 2F$wdo"0 5 ʹ3f"l.q堼+3%C_ᑂ3Qd8:Pc^7,ZE{BH/ԑ[˃RRMD CDڇ[vQ_u:ǁ9Z0R2xddJ <(xL%m,(U#ō%ꟸYHOdžۖA`?jyެ?XAw$y9 1)oPB"c E^,_ߎ}4=ѴFb,JՔ:ךi!@IYm(`L (s$ùFeTv#g:"gÅʬ r4AiS\"䣾E8bWDoe9\@ C*4#z;f\LAGlYMunn׳IS"зvnW2m,w+ U fjz`a!"e4fd\2_ݜ|ʥb`2Icj깉x+]`YJ=(t<{R2rJiLq:kCn4 ۟,e縣狻!0_₾mυbL~N*uj} "ٮWjTэ0J('욜]*9ur۳zY6%A#yDgc'ru.DWRvMl]glx6EӢY7V9bFuݍ:2W+$qf4 T#CfE|9ue}6F^78A=W6_nPĨgy䧦e67ms eH\+ueש]2yL̇8͌UhۙvF/\pcpHpc$jE.p/j (7[;gY/=4lf~:g6i JCAV9B\բouƌVƨxM44Tbm7[nwQz_%0.| dk_YfQ_1T^-ɴ)mߝ1Y|ë7ZWͫolM©īqaᇜa ǐ 'qԮ1: DRKᾨ[{GS︗>k5r$oh^h^MBr"DDAmx5] ܶ=s#^[FɱyŸ.߷_5wiAEw&`Bͯgv3-;|~ӾkWvGZ -HN~E\_y҃J$r~կq\b#u&RWL8)cBymG9B=zd\+Q(gQRDrDkw0`%ѨhuBK)8PfB84B\Uy8/oIl|Ծ5xmSW.&: m`^J4MD -rh)4<1n#Ta׍oHJ`vg5)v=moy&sE㺍?/F2R& cWC(TQ+\Jɲ+_hՕF!P U' Z*rfu J*%P e@sJ+c 8ƕH &()ڎ Aznl $ TR$2A3⩣FksN?$[K'-EnZCSᇛCכƵ%˜"!.B,C^ L"&"8#sl6*/l$Ǯ]7(A3i#z /e/^xi%ZƩQ"fN*&Q頩DBvR*4 œ><5&jlBpw;\%[IsvͭYuϐ S6*u3ܺj[os\tfGdmͦClYay66lgWZnd1{AZ۞];:6^q/Ψ&,Xs>y}C 6~jSZ^gIPhnn.V3dpl2T^'lExj ;I 3* | Mn.ZH.yxUT !Aj6/ɥˏ@`MB6]m^kњBg 3ŸGn|~wI✻ !U9̦zʂʖXf"E*RHxb*'T+s*GR_MXd9\*$TF ̊hX]Ys#7+~ڝpuF"h{f'3k<mTw{7Q$%bQU(Z-@ Q@&L#Jdx"5WF+gj-?vMP:MhrR)"F\r%qU5Gox*TBgxzJ J#Gqo}ӛc߸qӯ]ǫj@-a"hIjvOoN@^T8]bI=`N V$A g,AI&[O/k^.z#}jƛwon~6yڐϪAӑssi6!m}A. >󼐋FP2~*9tk<ϵ`t6/BT9{c7Ú7%Lu}&Yay^S?})/Xf2J1)+XQ4ypZsGcŲ.xDGZU|dۼ>8(esDQʦۼ9p:#IFL}Xf/syUrB yrN8w@<]hXCJ'  ڀ α}ZJ3F&Gʴge'`ɓG'42@Nj&M"R-CJ$ O9-YkgW5 ViFRgs,%KHi#JNPcb)iCL c.s:h>;'c-Bv%"+-PabRr >0Y" YڃZa6NOZ+[xJ=Ũ!ĐwD 3í۬* G+㨊'^Hfw9e!%@K,0c-kIJrPYNnjBCmYDRW Ee>2-ob"`C ,2"gv6Gţ/[DⓍ%uDR&J:h=UcvȾa[r7hN JMUFP%!K/- 3=MV~5jỆ[rI1RZRB(US+J >Z9"m-rv{qtr<)H1LڜIr^בW2G\fAozйrؤWv(}P-EhB7M;43!1!UXxT&\`>lv.#2aP<;踨ܛQDith>b4bΞDNz!( iel.HEdD$RC~Ok&Ul[A9́` +ʹ<&7FUX=z,޵ɨ,HqEl5}L)=K]:/Ҷ3[n/- iڿ/[nnJ7-JYnLZRB@'p7J+J.J.w7ZqfotY_.K Dڳ`hHd tMƂ4nR+PV{$)K(w " _dwhx x ?9Up4`rf+Tt;tň7>- dՍlk7W[zfUu׶jA=IZz..\]kG.Fj< YVWY[L`gY0 ZQmKQ3)?f2G>-pD{̹`$5M!H)%CtV)K>&GxӢTdF8%JXS l l]'ޅj<}lfQ:ߍ' "nO˫Mof,٪κ-xw( i^R@QYN:Ռ%h$(͹LL:{MBK$:8p}>C I8bX^Z#nވ^OFYnzl_\#FA-zEY>fe+ ]E\ʲ}BʲƲlZKnL놰l-v֙^s}{_ޮPxG wf>R[}<`Ddp.Jޠ=Rk  ʇw3_}(`"Kf~`l7_o6>?76}.΀dzxfC%0:g$ 8#}4N!,1ওe/YN0*9@6L giCEN1'c "):9sLUY)ȣ`%)@Z)41F;A|\R Rd mF>NlA%ֲ,)!pC Yjj1I ?6F-x:KqYF,G#S"Y< ED8OOK&\ vh2;i=yiG8 XS`LQ9fKumQN<Qh^`I2I=%O. ʾ?[%dzz,QEp:mhýwQ&bu".;8i?J$kzZ4 P]D RAdKx !^xhsI(f/^!}V !(5a;FIxL)!1q:➣U;)PQ T ~lq,q\isA[qoR[z9L^)1;w8pdӰGM;L-D2I/˼{wvwg!IT(o; nQ\]һ{nv}5 A7z1x|1zمhG; 1ǵmߴxБ`׻fOJGqQvԃ䛦KBCPOwVZj[]]ڦ.ݩԹQ;uۚzvMSw{M\u&Sa qbo[-p%z[5,aM㓛 =x ^Iƌk?6}k@#^q:;;-MO~=/jذ37i_%c2 iN~T녺އ~}uO%QJB W*gPtLJpt7O~/~e#_'+&'!)/npʎ1ͨF;mu|yJ2(cqR67~♸W=4z]\ӕp Oi?ݕG\ )涗 [$\$xh>xm$tq1Z.C wt/}o-i'3zdͰ/u!I__.ןACrh*xҢYwCwH#;{1Y"fK.<3c:KqfmKڥtI--,<{f}ls.7 BТIӤ:Y Z{{_E1"mӺby? .i b.Da onlԻ.%: RdEHZȳאJNe¤=[8)$Ez[gC`/ϡ%<^9|探30C2GTĚG A`)z,@R@YuBWv`gJ,]ew5VŸ΅{!)ZąG/9ٞ hMQ* n'VDUV#pdc;2KQ~tNb-헼# $F 비Rk˅Oڳe2ăn5Ohc~`rP{#DB0*Klr0]*GexF}LAG?^y^FlcRPx&F KH@ŤL'๳Ȕ;Uy#'Pb; rC3]@-DK=nEqiA:RD7jklߨJYNtl9"BKlk|oج:Yâdgu6|B֩ufY}rkƟ:ϟJERy{@I윷#3*)2! ]yo8N [O#vHR9ENېEZƠSV A2 M(ږ]2UZqƾP, U' omz񚈠a};OA~ FjKlAR@;@V+FXLk#Vq&Okt1X),WV%6 cVl9GbI%wEMP u;MDnd1t5rKl7Xvkܱ6-KmInx%r ?Ha,s,$ GTްWBr.&T;-8jўE&D%T{e~`]H"JuTֶ,[#g<;]G#=%, akq߬D|)0IMbbIF#Đ   ӪD4p)qIR Ӕ keX+fpO-K.2֡\ +ir:MҔ&'h'#ZO6nNpxJ0ʨWRʼnh]uST&YaE[:T]0XTq;^^3v սhiA;O"A Y4:!SFFT\ڔ *g*<&ξV;o"/b`z=E7 #\զnOL?kyB>XЛՍS+S0G`t-oJw,6$ -=gBH%I, A G 1GbTK5uf/i\hyJkBU "TQi AYe"ѣ2lk[}HT(6FƑʙ&k{(C(<\@R0$mWoàXRUOaF)&Awk:Y\,wz7x,o;-RI,8+%8o"r 0koB9.o!mǣ_~c>&KŊ੒*%X\iWIi@e*iQ @ᎈ7BP\n/x}~*#B}nlfUϽYwanMߑ-2{d$HGs;GaBE![pQ cgav軣驎6Q*HF3Djjc|@11p ܧ\H% 7ialTHnF.)$$p"u#6%E&BL>=;.>/sB}sAs_?(u3 x[e3NXeZ|UTqʪ~wKU> [Y1c?k|ԾC<]L=S  k9 m^J4$9շ^(1meΔͯa"웍I!d9-l-{t)|1dQLMVMq(dvtTRॳD R4L -^t(# YuQ\kZJ DAsc(S@ڽcBeLXό$!JʘA&$SZDSE ֶ3Ev# ->7CW냇ڵ穑" !n#XlC^ |&iNۈ6Bc.x2:RCTDi^@x);xyJ,e %q@IEqQPz&" IӼF]:l=+}̚uώ-]Nݻn/ȍU9PVmBν^|oRQb[ϭ˦q:ՖNmD٬;Ė%߽MY;J067_9]x~Uw-Kd7 ֜t5/S_my~,kͭ͟G7W\ʕ7vXIqf΋b3 5[ GuҘ0F'aT!NRD950Yk>o"dhJg)Ab̈+>Z!úZ#r]񩇽LkknVe*O'w4Tls6UV Ę""J&)I5(`׍v2|.X qR3x\N*dc^yɢh;WEfDA$' JNDy "o=Ͼ{6=-m_x #Ny9FGr/1ؘHWsqpCԉDcv2Qrks&9yyLɜ gIeN1:W]}B434!XR@ؒg# H`T}& j}p49˝j_KOt6꘳B&᲋t`!In\׉FK.=msï@[$A]|ʆrs3_y2rqxz:, }5I4 ksdCS/tT  Uc!`=1͠96.ՔVWװEoTv1vݘFK'}=#[@X.vʱqG–O?nWmHvu1 H,61KG?_57KWch. ֟Aa1z8JI!b/lU)"\]uWkS)L3uz}0ux?k9//T{hfg$ v_Wg_+dIbS5vXQDaS֮ȗf"^b4F|>M-ynBF8L|nv7^g6Z1R_׌1h.$6C5q=#h~~ n٧i0vt/bsZ+0W^zq.ZOr zJeRL%M‘. UWi+5p8g@4 IaЁ#R2Y Zzx`S{ yd*$p!g*Abp.IbC+2-6m07!My֐Ύr3ra5)\fg㏝d.Oey_rcgGy;,lx0t;kǵ'[֪?'-@dWљZZyY.UMdAcA"xlHK6Qr'TWNT"_%ͭ3'<*6o%M)O־Dg[dۀƴ' Z@A9< Z}v牠euCG*JjW޼RVuhj;t Iw+{+nWtk_UJ (C!,&@"r_NX#s.pЙ+lрJq,Q&(%sBXqnI>单\ѿmk:"C9Zqqm)%Ǟ{ѭruGP?.7!J(aXy ,H.zN1&qnk;8W:!1+Wΐgm@EYr҅$>sm$zHeZ3ϲVHa\) G'4 e"+ LEP| (we>km:[YW'6sYCMJ-G2i`dDɠP"Ԉ 3GNQW1e1(|Au,& \|)?['AX 9fK*-wRiCdDjIqArwGJ '1OQA ,P$Q:@`g nN m'8ZcSl4)gVɍIqyǣ |_3҄3dR%Ygo-C6Ǵ:FҼzdi4Ae&*eAqfL$0F:3;nk"Bs;!(|ȮgZfymPŤhI (5WNrN#dcc2;Z]ö%A`&5&G?d9tk5r5an9ؖn0 SRh$y+BA{WQDCP6$ƒn %L( Q tV%gJLD03%ްl%g|,Q?7%lԻMn^9\BciKuӯ8Ⱥ-WA, Z i@KyˁY3+2JPsÙԙl7Gct\"t8eA<уw9 [+&㉩rdm+Mv^> N[<{+[*)fD% {ϨT(ͳQsQfV0xu<ƙ aËA:k #e&- e&j9֊AHYCf&$o$w3{PX"+1-d>h$ьsh-Լ|Z<~_0tS(vt؜GV\v ~gzq9ާʦ%OlCY8$81'dr@# %>YcM'D"Z޳xN0(9@6L sFE^1/g.Hd4)s 2]zes41%e1ē`>p4RhbN).p`Rdm'(jm:[l )t`YP=-'[S Kw{kdM3l /5X>IpF,Yfe%jFhKALVU1h}"IָRPu3MNR'z֊N|iGx XS` ʤ{8i?H^`gly\5f- Pm(j['H a-6xaYa|/^M1BV !(3&M؎s$L)fu'.w @T3.'LRDt"(ms,q|Ҹ:|o>I\7shqnR c:ԟpǎ_ oaj֬[Ob`eֽٹ{M/q{sn.]{7nv{@n|͍یaŽیW} G9mhmft$IH;7xN_׃^Qsy yݙ#۰ߋnZ׷b#vvҮRۭy:7jWpۺݱGGۭ::ցerbȯ i kv=/ ÿyxyx~Sw ^=a|1g{ǿ Ϛ>H55˦rkV oKFKRs <89\rle=T;Q6$U2QO}ln3f`uU:AL,.On3xNImׇDӯހIc_pB 2lt~42*cV67SVLܫnyECW=h)Sڿiˡ46 ?n{>(+%㕣\7k8cާT=;> \򃳟~O t\澷 }8>~ [8c՛4ACh*xup֬xwH#;{>Y{,fC5-OXq~J7fA2)f>Nq Ψ@{^#WV$&" V[“%]J. tFY]qHSa$Rq%x:gor-+oFoܯ;:K doR^J-zyjBʱY6򃪺״ȍ죿i]ӽy3x7^W9Xo `*3TMh |^3˄\jwto`'| Ym ^n6OO&҆&wmK &xCG}/AIs(;J}gyD#iYǯjjnfkhxXҳ_WnCC^4Dk 3县n)>I1Q! >!:݂o5w3җu@=`0w( Z_:thQjci&zG1Ns]0aZ5qK'e'!+W]laTQวG9~J2hTkmǥGHD[X?/-C>ibΩPv b08>;Z}8s=܉`?M ͸8/ovqv]j9@!(ɱqDQw"µ_ BXfdp9?M6JWsEśQ*2 U1N4QH,3X0Q#R/~_~yNVQIo[u!2AwM%%<)Li.5%[<[N?% #iOM{ɱ.QK0QX"M16㎶<< Ƙct2%e!Ȗ$sԒoDpZvͳCT]-Hue˅KM|_>&a4ΜK^,{Ya*5%4۴=f7.if\65| J)6M hGlzsҥvkvq-Oې{m=d]8֛5{ې.|ƦΣ6Ynt.]χRL2Um}ЏzgeqG>Y[ f2EdG._}0nIh*nbXm-"v*5\A4C"]߂D j+;Qh|N]fHM T:޸$cKשwaF]ƫ qQYWgsEQS7qnjCvwt3EA^[)QAl1wmQ+R2=L M={')o͠NcոCuk{K@0Q%Cj_fU{N-?`FǺz\WyGi#eP<&d 3UmVB(ewQZJ:r$2&! k:aj3ւg2b Dk45[!-dNNj;E)bsppvދ|_Fvܕy>z=1Fwy, I*X51 KgyK,,Q;|0a&rroDeflϨ.?e O9MtݩdF'f')jxtM)iowciM^ v7mSŪ[֟?.JB48O @Q vbQ\QAYH#Λ7O(֢VГ=)Q[W12#XP1g&-MgḋUlq,TeeѳOo2ޒyXroﶺtPmAiѧU5}D!{, VqTFIh)-0F2ʣ)U  ̄* c= JؤB:`^e) Ӂi]IYbg~mGCdbFJmYj^jv'KN$LCHpm>2%>by0ޠ\0Nf iMoBdXю  D9d|`#Fud٦_F,)U'#sǁMM"ȿ_DTi4: " ^5)PҞhIrC:D IqbKFb3[XA`Ir0KCFbR{fm:%b]Xլ):Qɡrd\E$Ϛ&'3O >;&'fqcʕp; @}lɑa6KL/ DafiPQDg. <9|g 6.0u&A I@X TM6]`6*Gmy5ly')KbZO#,4~]븢}c.6`vPEM3NU#|/a7߷)ogI&Aж-"j,@ϧ:>\ ~b\`SY{IW ig_;z7 VBvWkS qjbځGؤ !Od;(Zr&Tf5pc4$>l'.O")#Sb!%Fy$\{0tM#$ա#H00'#7Pʍ"VFudXDcZ)"VSF4IrMgpHrEBR4 YmHx U#,^Ģb> qL79 Eh.+ 1.^_(=Nh 5R{¼-z{ ;ujjf_nǿ L{䬠ؕL9mo<}#;4G`t7t?F__36@KW,F R2yB0{U}oSIMI*^"dJE3Z"'PN٫$ɷ`ﰗGAwq$_ضU,E.E.zlŲHaI#9+Rv9ڤ#F_קVO/H-8CܖK)jAѾ)Uj=)D49ZL٩JgbMZLtV:SIa `ԛbQSo7ńSo7ń#-A,vx<-{A,EIZB,_w*ltSoOX0=a{'8 . :Dlkse[Yt !UV2:dc],!+0 `(K z6/SЋZ<4֬%o*q9Z"GJPr2**]'P)F~~*u̘DI6V9lIޫlOVnyXQumR}($&'B%O.$+N~ޟΎCn=ʏd'N)nj"F;-LʈIJBl0.*EBз%˜UjJED h1=C֗JI%ԊsqLޖrJ7[M2--ܩ-j! udǟ替0Gb Z!)NQ+@.=[sE \=F#19gŢII̦q>;LT|N[³>k9[tzR{CK:jd'{{}JF7!l*,EL(j-qdzP ]aǜY!xٰ[SXeY8rcS1jT=출ԯ*aK>Z&8Y Oi^^|1oQ^{dxk@ECMw`yߘ-h] *ԨR@Y%笊5/PEs.' /tlkMg{h;0a{~%MُϔWǭ': &:MCyӣSϏ|(ߝ ߞ0Bv\jAvmS'غjԮwg­WuMUu|8O1g秿}m0_\N8&dmEB*) j<< ȍ*ME>W վR71Jn R^?b )PW4ΙR)hA+2+4&.Mv%Q-.{.A}vraɅAvwan}Ǫ<I  !:(1{تUJoV=}n'S\ۭ[DVR1zz`1R$ WԤK }٪;duєEdK`H}TJbNrP% {3Qj5}1x?>gXL*j"lTqISJSR*e6??ik\[ӣIkv*֑c E/D\Pf$R "{"mmI$u"  T,6 p֜u5kpw.ټ*l~Uda\^U֏mm2vb?-st2o.n6:̏^ Uؤ'?nZOjg.j/MW??E8߃wMh"?Y<)ql_hM;mvfb'3uahNB 9;^꽚сvcհqƿܞQDUgLGe[5_N?=d9M]FRo۠f~g1m5fwlr^1 c)rL bnP#7f*^n#Z/|yY/J Z??"JG[b6y>F1߮K}B|q]EX ڄ"B..5@UۺI8MdcNEyk݋|1[1蔨`g(FvoɿiSN Ny>n>,kftI3Xl^RTv(ѦZkԜNlh'2Q8`&lŀ gc(;ﭠ Uh4D*'a`7nmP䢪WOT\m*C !SeJpkoR^y#nX}BOHlk/v Eύ^7ڬV?.b}%r|L f} b?9_e55 wΧ=~u߿XT۵!߿Q>cʭh1N?>ifcx*g򜟶\?|jv+憧_+! ы"%YfQ=r8Ջ{|#|v?ؾqr. ] 6]ѕ*UL6,u~\|4b呟F Vjut`"[{Oxdq/.&SRɲ͡@ 0%hˎ7 zn΀ 8(NkG >lY H,aiOw)`ߦS_|~p>r͏s6}>0r~Hy uLsb6_2+n 2zb%%)PPxDR1n( qJ X _X2Feldl}NG~!k7lӺ+u՗Ag@(",Φi(hi7dYJr]R=SA"xjHO :e*m4<|髍TƑQv&H6DM!6&O_Btkҝ/'ș \݁&ާO :@A= Z}3A\ F 7rRgj$&24y4K g;Ũ?*U*A:!Ҕ':x) T OL3eb9/C7jSUBe׈ҹnT\ц_Z=bL\X۳yI򋨄2Nhz!LDH<0*ɳBcrlˆnDPW\{ /X̏:T\,rT T\A1v_Z 9(Et~w%v{ai0eڿ<.)7⺴(El۴Y˭q/ϤysA-d{C\C*kf78XjXbU)OQRTϛSσJ&^D*: ~~ݗR7.z ,+d.ެ}@g^Rkm>CI)-Mm`*6U L"@'`mZ΅wY_We_>{ ]MvVf tӺBً| &I+1Z͉& dTm8- sRE )p YaLI.R3HRt, U^Wo@s~׾n]VVuvboY>e)-`!YIdG74$1+H`%0fe"ʅ\ȽJHg^a&?@}eov~ nAaV>:D#"Fc!b JB3 vsu(2L>=RֵiP{mA>*|J&Kq2IČrhE)~e<(υ P2/X bxUARꁤdF&QCbAZA,w J๬ DTĕAHJk$ 'gJZrv\b%+]tF.X;5^L J, Orw-NYw`)ue>2ӗ~bDPr4c Q<͢m@MDd/1;~OO7g8iU={ifeB(cDRaLph0[$-(ˮ{ͳp~%ٹ ZPϬ9^@l$- "QIE@* <#rA"2h2I7 [w/. }%7En? #^kˮ ߃ȎRJ< phuŰoJ*Wֹ$eQ,M1HvAm&D(-jj|JGB!Z@z4SB9˙g"pt=n `\d\]QnVQ 9og^)LW"Dv; 59`]Ѓ*nrKdzY,:4ݫכtdߝ?2%l0ð;ߤw!]m{]g_=h 0G~:_URhT9UM㖎sMOrGҘ>VJ)O{DW4,dWHdkf.u[Roݒybm+GQr5mؼ.׾ıE iJagfRGgvK\~썎nD{yB^cF{ї?jGUzqz )hkvr\fؠS[w]e3uןq<2(|E/n8GQr])RL^ H($_iTp>zó/[}GY:ߖ1}`MZPIGJձ2aӣk# ["~B?zS?m"B)jK*NŕDZЄ+t{7\/U8{7vies7nbmy* P=γ`BB:>9>2\я , k:.z3ۊž>47EVBIeߛsObћV^F~u|wk o]7Ble|;@zA;䡻Wy6ɴ%R9L&xIjڔȥGi,67wPT7p;6LytwFfL& J5ֳu:5a6=f}7&Uohv Xuqo~8bۋff`}l@CV)(%49 1TFL* XgMGdrJX )r >ƅ=dJJ;3(φ-aWWJ`CYc8ʳObPF&$jeJ;Wz(́>h:7n쒍ovcaosC\jSkSk,J9 DϔHCQi4ouRBrzk0Ù.Dx] aD Ap/$K5''\]KՊ=Rh>ل@& %2dӔY؀(oՌ$-%o{f!rkBNH BbȾ0iqg\EAuq57oV~M# ٝ: ))-1Hcur-p}Tqћ/5U 䫕~40l>ʘV LYpH7Q8mDUA'=(ՊcA~ZZ$i] Z#gwL62)sY9 b ^ R7@:RT.=R{=Hh%H!ŨL/E]@p]b"iێ{Gmd'))YZaBuIƈʀjb̅ 7P!"e8`t<j[#ҎVjqi}9HWm>8ƕNPJ,Ɉ$S")02Ra#!Id-no[ږ$ 3ncbLQfĘIzepmxCD'“&yg-Q:dWm/@vUbh2 .PsW6df+ #7AeԹ[I R% )N$✕)F\X?5NPHCRi87L@yq+QY\PKg*r> mRfsJR&%weqYW |Ižô00my0b84\lw7ܬUY$Ւu2sQSI{k%xN4dlc9\2'%d >FoaJ>b h-N V\j() AJTm%@uP< *}rtU2Mt%C RbtLECMZL -tF\8g4E"4iՠ,C Y0T~cм&>y _V 5ȤuՅ@vTm^ ,:?jPJ4%;+QMzC*KtK|̡cLGu} $$LES{XRr*a3m>%tB;뒄 X}t kR([|P!D;f bro鎠gc0-]!]c~?XTЙ$ fj2RAJ+1A2~ȃR%m^}݅r舌kUPaRB]$1#dlD(!hvĪMPk>Zl?vXIWH'Y g#MGhtf%ݢZMnU{9%D>2ԿIIX@it_tВ&ƐZ*mmz̃vW7͋@ax0ަ8]ۆsu&j$P`0u1vtFOB=V`&MNd2YYۆaMEѽT$4Id057v('g #6l}Ie t  <%*`Crh[SP.Oh7"2UwS[4^Or%5t  Z`ѐLQ%fTi SRC֢뛶Q3l]`Z N;JPD4cҠOn A~?E bZ0z`R;JQ[TcD&7cGQ3!1RuJ` @ QYj0Q3z hNmS; 27 _HkҪUg($u21KCUZ?d?uN^YR8N| *9[pbxZ*;k*>Y[tڠ@~IVT@; Z6i =W&Ex4CSQ3A%#pbe6&TS.\/H`PrAuƒf5ԤU|I负b驋l$Tip/hm.?!t+<!GTQ~=Ml͋j/fDULfJǡKwѻL>m}7y 3^ՐHJI# ryٴNOI,1!ĄB Ri t'Gv g]NBq O $1 vs9=] 9>9>а@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v=C'$ d6p@86'2XNB@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; Nrx{HN ϰdN u`@@͓w$;(Gv@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b';H @R8:=O D(b'3tI1@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; |@7NՏ)x}}~@7pޥz|~VW@Jq c<N1. O޸q9ޟzgm$KaqGbdA̞̾ک&c%Y!rLOA}rQ(ڶߑ3iOTZ8l@/;h/v]]'*p]y͋{>GWzYqg l1G/j]iz5G?-/֯׭<0UhC? iitk oKt^H̋Y[bzҰL!lL/h_Wg&msb@trۼPMg8[keքRk+rf]  ݀Bw 4~4I[Yrjj0Z!zkBM<ɓL=hw3˶7CKh`C9b钉QٷpꣴP'Qڃn8hBa6B+ PJӣO~B<3+0 !XnBW@k~tgHW_+g=BU. ]vy HW)Q ]9c?uœ+tR3]=Cϰ|TtEp$VOf9Umou:}[) [+WK{qv%ZA\ (G/U d\kuE+>J2!hSjBr%oZ9>3.AU߽OA}suP~qNYet< \`Zq0 ~ aZVC ؃+ Pj~?tn9XX?sz'/ }d>] eOL];ЕbKK>I)ubĉPZZ5;Ni^˃iBkSi J2M?GZk}ίnyQ|<;⸻mu}TV fB.bcx;ۗGϗD`OjGNt|6s\v\"|p#u:F#YҾZ6F.R\ʹRvjAV`ji*/tpNc\T"o=VΞB?}KQ9rI?z0tOb_E3)h1r)a9>?틂/ɴa\m0CPZ#59-&gn=gA]OnAPs\-EsZ/ J/Xs>Cͩ|;]{U x 75e0jMF8JTe~|t蜄#Wu&`0C&[3 KQT[r]o}.ߐZ<:;ۮ"X?jȋ:ww {;"O/s9Չ!}FlcNק@EsfSߖ׭^&?٤yGnw>?'/^TSk P62.4QA&IʎFtw fЛ?f@[0;gp((h^%{jġx/azb (g;b5TƯ^PV_ˏ?Hgpztzsiͩs )C==ɯyq=?~!?|F󃇧>~_<>́~\̍!H[m:FDVX >,r-Q*"nhuel)Rj̝,sWo~yDMҎ`-SG{t))uZMJtKLMgS@7aMy6ߟRW7*t%.eu%h,1lmeaDa|˱e6S/_-xjTVʄS:Fks~VYDJ.2Ku`5*g'qIՏ׋˯n-p,;rsv Yt~r][o+Gr+|CxTac;c'zTKC\Ήq{DI% Cs35U_UޕJ3!O1bi8Q[d57n\ M8ZQS]6uzcVr~9x?LBp&foYoS1 W<.]uz:ގw٦eV1%`E,23 kItǙVGcd1 NI[{rUE ==kE>ݲbtE,aYLEΖۛPXhCӲt5I߈Tj1IITncomQAEjt\zp^0n0BH=h@Q`jÖMV4|RHhaA ]\pUI3(T Ii-!)pD!)ЅG!u,dT.Ɠ6 K$c/Yl7$%'[ 2#:8i\ `g9!Z4YJ$<&[Ʃe)NCdЏWcȏ0p+eH[Gf L(!de!-W*~\`_۪]N(,I0j u [K E1h2Xm.+  Lzsg%91mdfK43}eؤp(}Q-UhBMkWDb6ȋWX+s`4Y@"TXώH0t\T-W5cT%9c1e9kmȚ.H{FMq$6XJ_DvNu2vD2V^C~/!k?~Ŏ&U|A;e܁KuwrQ"M6.VJr @iB\9omT*V\4{exA$d{h=Z;vw[Nqσ |򛭛yиά@K?>2øz?^i Z\|ן0 l+Hk=uB+> Ӄ"+ր_(NIL!Fx .ں9vБtãxVgZyPQӕVVYb]]sA"ܐ>$2.$,S(dJk.I .)n> #MJq!Q:eBQ.'Ж5*iT1r.9@G筈,Bv%<"&X0lbg7n7c{p^Ǡcyrf z_[,ڡ%\Mgp]V9 pIcN8I{pGJֺ\fa\w ûR [%=5S©4")s :u\Yprڹ,uL fx|eS B.!c2i~Z%;X5|A#6ߦH{3EtZ<`K gxU#~W<ʒeFRVNKLFX4>&֕@U˱ĩR:ӎMfS=I$Q'(ⳗie㫫wLB_ ]h 6CCۈz(j۲2 %Ԇ/ct?!FC,4\ kb>z+FJhDH>|!^g_uzP YJ=gUt8*̻y%~<,J 3/1H<s=?,'Z6pҶ})q?$;\b 㨞V˗Zs_Iu&nёIzx?<4We!4Vh[^7SPt5n>&ac/e(+=H=ަdխCV~UB;]IP)#p!}N" [n/cjwϤdٗ:~A}/_.l4qXUqp1D!ѻ!M:ߗ;+ w'.y-hlOۿ?V{M{'Y.o]S>)_w x0(1#rHPxÕ)~"-mh>S] #ȵ0cW}7P0kE[7ߝ [ywtS A&2sYaMBј.X6M촾E1-ַx4Yq"t^}@[vy%;;=9ZX%"*%AY,8 +f}.{!]`*gSFɌBʝ7{{G-F& Qh@j͈lEp:'I v*aْ0j|ZQzc>8?Q5uwlgjԃ\q]81hh:vf]_vǁKR!M17ɤE{4@j6Mx EUVxLخi4%O;m‹Hk5Dڭ!=>{mڅSG׎heo{VATgnD`^M7B{5݈Zձw#*T u#z݈H1O{f H?Y)9[0d`\f%;]9rU|>(Lg޸wWAnQkl{6b`0ye6f/d:mHr@OT>wy?SHtZ?іdS12UV]EGȱkNt80'8l/Zc:̅ Q7mZ[ALx+Ha,q>٨Ȑlq%#jU?:XKkR!8\d3˜; in95rv ۨ)W v$c4}d0!w I^2W(Zrˌ+]'41hڭ5I@vY)G@BbTdA` eAm킊ga]MD%OJ[cfjOf2G4M>קUɢlmA<9ݰb koC}:jd0_KhlZy{hl0\NBЂ1InbtT=rc۩)U6C6ZeRvA$ `\KϹII[#gfl [}umYNMeA Qu>Omq|T~0]ؤ*ql5I޵#/7IVe`; nwn?.Hr2[l=,;rR@[bdՏUbcUPiCF42$RtJ:R !cE{&(zQ u&ێT*YЦen݈ǣfb jmmڽ{E V)ixu֜9ρxfBf..x#9L!G$MX J$a"ő0>p>fY :iZٍQTd-8kM6#kAD J El1 ޒ5@\ol<{;i[EDk8FD6+!j>L*)!@T:2"FnDlR;O8]=%_gk\r(.q616BdHI/ d>ҏR")z/>V5{qYmhfzW_ݹSnW u[g6˔ѥ-r*D+k &G^4Mhx'ޓ>fK|黢TK_ t'xʟ-t\fv ^zX\O<ϟu;mq{ \bJA̤)` Xf.%BPLlFsL{.һ=ބ9 # A7a|_ddO3YoEb<8kF)Iـy6^JJ't _nw0`X5F"jLs`sNs)ς!ZoiyLA u XebRkRBHolj27OZ=eHfXoq a-ad@83n0'_hTx؍RvòЦr荔*(L z&8,cZr$e ]7Yi+:}droۦF7Onφ@ JBk 96dFH a,`\D$G"/'AcaO4O๬k1<-a}{#fOzQ),X:ǎjj-=_Fߚ^jkKHJl #TV&6DRt6EC.}ޢ0Һm/H AUJ#KRSE/<-\rT0QR9x&U dL9$p,rh^f66}Yh63Ll*\N mqitiz:'ϭۤ/V[ul/MU G_j%@ ݜ#\ l zxI{[5ށ2&-YHV=0Ν:dnFřX|/iV?7 SND]}6‡ҨڅMer fPpNuoMp-_pQŁVp]₿m(pG)VLD~ɋC gכ2Ӎ 80qv.ttFYzȲ9X]`,yJf*&=LnI^"7~w VzWP"f SpvHB/lx^5EӢYs6& w7Z99 abȹw3+ُ/nfs74:iЋWF7b`a)P"f W_i- ]joYokZxtcJ|oUv٬gc+nbDI$0Ud_d+dlGG]DMHohxbKpYE)f8YCkJɀ)eud;|)+RnsI4>m!/2wPP¡ӎ=Mpo%á%8N,Jb Xծ jW5ԫvY_T k~g9䷏mHY"$kÔlڃpLh-YO)luȋ䑵V=9EJH&',qΉKs+Ůa*i&ǒFتU5fP!F)c)FX^l %>sD쐻 ZZ#gUG8PPaB<4է ^fmwM`iԘ+poWS}/#LŇeBUX9`p)s֣ RY?bGw^<8ʽͮڸ/lbn| 7H9G+T/U^zpƝIgWZ(% 2  DV,O:vzuluŲ:n>fwt؅ﺽ$Dn%mZ6an+KȠ!oQQSZ7/yt;:]q|f! nޯzmE=xKoߎy߫\tE{ ϓި\MXr˦^p2 6/ikuϏRczHZEDrAv%P+G*uqL\u :WDb+pU'WJ5•1.u"9U!WtF*":\*•-]2`١\UVSBz5:W38:\=`ԑq#$H7G?$ WաK%_X8mu@k"EBs0`8:  ?0~Ul4liqȚ3BĦpn4}qPx륹 WS7~(k|śQ<#4X8] ܖF.iV7OV/4bxiR)?ZT9e.1ہE;ġMJ悝jap}M-B/8Z Ȅoj5:jbQci,ko/QJ ֳ0?c3sl{ҬWF:^'luRdʂ~6 jmԒiUXXc t eCyw,B.BW,Bn*-Wh K \5v uZUR^!\tO^r#t 3pU5ከԬF t$dgઐ9+*{pU\b{z=pt u Ѯ *T|p1/ ~*\VU!bWJ3˸9u*T4\RudkTWU\:\*p $۝jB<8 D]_]GMYmn HMf65yII$ayI\8eƕ77?oҟ.yN,(%7ʹJ6Cɑב)I=SEҎeE> ߼?^sxb:Cyi78YS|]ufɻ͛]Tc)̪̓OYMICDE"܊1eQEoG *eAy?,U U*} (eAY΀nf@ik’xF_p7qh/N=~#ǚ5~^sYvOڒm%eK=쳜Knj`;uFU+GE*FU=7?7Vp8r͑E9e8*O p=\\m|W"2* RlFz|sɇ{z_j`fJ= jixF4 ˗{aIJ ,32B [)j4ςvKJ0<9e9&쁧{ȍ~;gH^n/@6Kv$AeFK~ŖdKa evȌFMͪ]Lbgz֚4#Pe2`j Jd*+N{慱W RPx6T<c?]y& Z2or<{7 2ZTI+ͬ{H `FO I0ͳ*nX~B5:+ 6̮pQ{AVK5xQy/ /~JiC} \eHa;FΑv0 ĮΌ?܌'%*j x>i 'g͹0Oĝϟ~mtur_ӧYj)V?̯7ܲ7'y}KЀtz +'zgʍ&'p?ZVǿ\μ4]F%cs,]__v{yُo!D%iζ ISbz?`\^?5W mpXsYWN. o頹b%@} nӴzSlC[w6zGq)(Օf'vq5,W`3_~> k{cg瓑z4,Sa 4˳F?\-]HҶc IIHvw;.SMSjQNwϟN54FLECd$Ϥd迾M 24W3@wE*3ƅ -=mSm/ H3o4_|SЊ .c?[7a9;^^eq%jEټ?'&֛}pIϽ+,23O?$SsBd4uX/^WQć~͊!"L̫7 0hsiE=tI~#i 6Cgx]<(gNEv/fA_zM_aϻ8CZ 0FלhNFisa 5\O<H6%ɁTËd<;4vwÈvީp6{҅/~>ox!R /hjxoݔwcYL&31BxFm .,;ul6 jt@bӷY(n.*Ry&MqS̢GQLeoq&R qtU:sSeS~r\G:ɬNYm`)j:-YiT x^W$0f㯴~kCͫʶ9y|9pX1q~2.RT9: Yvb>n[K/vO7oٕm* ʮUgч^Hom_2 o K/.Nz?uΦԋr~g$m:6Ih:,Fa?_/ 8}R̝,)GdP4}A}Fߗߡs9Ih07'HWD]Q;lp {6ɥh{2JekyZsg+Ip# bj-ehvrU!qp 0L#A*!.qaG;[∽؁85X\rI:!jH0 HJR(y4e4Y6}#ڪ?}XdbCNji|Cx4^e(9 b>GQ)զ@pH n*-o?C]˒ཙ̯|~omͼj{6席]B> DžfoU3cU^L5>@`UN*M2720הaMJxB}${Ή=2n`n+G=#:_ tIKJ P!d2!J'.ʘHs& G,JǸV5p7} ]>44h5j}7wgާ`ڠ4M]Y6+{L^o[G<0] vV_+ IפG৅J^f)z0fzy#Kq&*kPޝսy{"gsYc8#Mf6x{yxx 624%HfLcp du,8:J[nmb s BC4@А6G*MrDP{ib0&PX! ?'r@1iȖmGiJ!6x9=ge6#f$#`xRIPs1:48Q%*`\s*GT7Ep\eo=:9@MNOg 1B(2:HJ\m85^b45qמ~c9B[K0phn|"f]\YnʙRBf”ehH L|t!$gDG!D!ݑSH{ttGA!, 'UtΡMRc5z&$bI\d\B @Kԯ`8N.r\y*QTFn5JN26]qvd㞞pyV/o&d6@L6xl;=纃5` cWd%[W!E]~5lݥi[=<+_23uJ;[ϲ/O/tӚ&eر9z&Rمn~ьԐ@ 01!ZC̄ \0Bo>E%!]@k䙋K9&ld| f]y" L|F)2h IQү \Rdh3Ǟ"rIf'EW,(tH+N7/,̈́(F)-2fz1 0.ŃMI4XlQ T ۇyQF3YJ8 }>آ v`-灵uzK#>\FAluts@o}QEwGϳKiJ-{HEzyͧD 7,ZYPjEӍØs BJ*H7z!CY5n{u  t-/z"ewPBUmEޤ!/mKhcÓ'e"k6 S/J㥅<6!}wvg)8fu ,% e! K(UȄ`R :,nR”jo-w)E/9{ǥL\d3CGHO/tegjg>T#^?_Gĝ4/hxs3 YZ̀t鑄U5If`N{I d7,g˧P$F":'*zc p+j:->E[鼔ċ?Պ&U}&#)4^\yu\drڞlpm[뽶=& ^/Uޞ0i%>`$ av HM* &Bȭg;LgeNh$VEDJ#n5X{TF{M&h鵌Fh2q;f]@9̂-vl~4KXv5jVuVC%(aQ2XɘVCdy̜H䍚IdLU{6U\C&d%ɐ-Ap$E_YET'SUv{ؒE5b"֒=-`u-5;Z,-\!.f@3;@Az[C(c9č]ֆC4Rbg|ʤ 4JgY;OlWgړ]4s:jR]"bg7I<tT !qbRH#'"*0bC$"03iZ[EgaV}!T=< i\z2Mu7Ysz4s:'~t :y 7G4cH n ss|eatW 8 Z3OM`%. qh;7n/$Lupе`2*{0 7mR)O1 ;yCàC.omZ7Ժ}Һ=dK7m6Ao~mԲݻzyvyfy(]93";9-ϳݖ:p5ѥ^xh.t{yM{6jǤqϺOB[q.|_E%jYD$F*7,;K z~0Ks}wp#Ɫ;gr?Z6JJ#ʜ!X WIayy4h d ZDNZf2\Z3(+1H-E4![WB],ZP|Z|X}$Um8;Rkz,֮<|)dI?4ЛZWq~y7}-6@riCiY/iѹrR:hmD>DG WGft.\R2m;bM\NJ۝e{QDBI i02\.NܐF$$EHT#تG1c+PdOY6h08.e5OD/2#,$aOyϥYpL13d܀;\C4=Ҵ" YVX"&:"5KQ .}VI`d\Hޕ6$׿d]Y"C3*+Oxwv?:)Z @i4ߝ8RDMh.ƫW/2H Y~K;.: .UE%(/%,!=G;ٻsmPO. :xh93Z9hE-T1-\-N\' ࿖.~?^GBq#⫦΍iB8go}14yו 6gZh~yqHzf0 Aԫ^Dz5NɾnM.Fiڻ-i4u\4#w~WmǏt!>zRݍh9s'8N>f'w!nY"Z}ph0]}!8_#|JwܸdMuQs-1L_S2,ї'S 0oKY.p|gҔK6 ߭;p OȞ/|=QWrӦ2hFv  )7i[4Ųj$.ƳXkr r(N/(1tRiD%2n& ZpmMԓNd-{)B7|@yc,W'vW`|Thn{\4HI4zE!IŁV1T!"K!]N](S{@+aSգ6-π%}I`@!kBLgM&ح]|> @n#Ω7N \hEkN6e9@cЙS1~62x9yV:g'u|X/MDHtVVȋa\U3PJ"dˡe,s2+k(&KmbF"G ,r.iLh9up t>'V)ƹ.h yZ֦6D3E)YW4$xMŃo/Y/⣗Ca_C3P374QOvWw5}] ?eGA1闝SLp` \Nnp<qѮ+(q\!ѷ BH8kp:?jEu4\躣K/B5{.vt^ʏ ?Ç$Q2UtJyXsYPDH"T>$O&K&dND'S%1@R&8}ʆGeJEFoP$S٤o6e^#S74 'X [ {RAOٓ1ָA#J!!ĖWT 9eKC>ܭlttBvTJjY&m bʜT^XQ-}\ k3DmQgt hKf2CƜFd87Aym$W'6]18kġm:j9Yw4llwgRҝz# &ŭΡCA>B$4 8Ͻ5VR u Z(*{sbLjFp?Nqe9 NsymU%']]A_ŬTqY)CB 6LͤQ4dC;8k =Y8P<х¹eehɠ2d1\xh|S&/<~49>=_=xz_\R"lX^[Pf"Ӵ,#z^ Vgd ݵ¤Qt~*Ό:Ci4~5"(Lǎe5f"|:mvk# YԲ9)-Jޔx d"sgy%!9sI:IXuy6,䡣4<:!.^ΦwWKugz HMlӏTtO7y6ްR RrvUpU\3#tCR oxܐբՇܮB.ֱP2bO>z<(A%gT_AY9*/Yl6$%G_ 2# ;8i\@avZ< * "Y9 ްNm u*dtfx1d޲[nI.'&Bl9@D^?&-< _K~g!!5glP:JnL)ȣxO6h@ϷX|Á!e;nQ7KdX`:f!FnT #`|G2pM5A;U[k!P44pE{r%gt' 6Ec-c)l17O}ܢGn_ 0ʡFMwr\kC rj>Px]'Q]\BWTŎCPJ(.]!`ɺպtE(7kOWCWV)PCt5뎺"w-ȶ4+]!`Yg "ʵܳѪ%;]mxkWultZ~j;e lAWC7j8mlgn 'C]rA? wA-<-w:E%Ԁ2J_^L'ť ;g8J2{l%jWޡҘ.4t T]iBeiPZ!Ҵu_F5/3?Na:T'2q:bq <^ɠqURPQT*%E٩g`}>n4Nv!TEdQ)R *"xN(;u&:e(YY\JEaQ`}#\<zsViGCpyg<B{4Rޣ9@Ft՝+kXg v"_9DRD!l<\i̶v"C+ЏīQBbAS+th=w"tut, ~"u.p"P>hn]!`.WWұǻ ZN# k;]J "]9v+ldwv .5w =]=z}̬f;[LW5{VW[U|Oj;e;z =]=9D+՝օuRRj{灿-m%w}r>YU#e+ىŠ(e0]¸AxmmeJ*[ހA {A0!@5tF7\˻v(ua]rs+lW *m+BDOWHW)݂=;BwBv-(z0]!`CWBW ;w4e t+UStpw]+BdPjҕaV;!BFբ+tEhMsЕ.-Sv:CW :vEhe %螮Ctunw kBRt|ܳ1s lX=^{;fj+@v(yvtezzhs!d+([ BWVTC+˶֓;|(1(ksRydv[YZkz^ ]h1<T7LA_3ZY}.h٪73^ܡ-}OQh9O 0mV>=9B)E''Vw33tEp ]Iɘ`m+B +%r]+ +v+DkxP*Hc]ཟO qAZ(B)z:DҊY:DWتuQWqv"BtutePYs!"Н)kXWZh;]J{:@h+캳3p]+BdUۡ4}HAҕq2EG5}'1`7AkO|.1.iۣUJf=/ ߾4:}5W7n:ѷRA_|ݴž0z3КG3d\_<_7 @ˆI烯rtrt'l<.MڬS7x TTYT[?vT[KCBZ>GqPP,?{WɑeJc}#ݙWvWcٞu!>44tw5}ϭ&T)tVV{nDdU/_3>ҵ o}LW-qhݖ/&n9vF 9KLC%Z-h~mnSkߜjTNSߜ*m_}g0هˣUpҗi9?;>c =km4]q|cɏu9~ޢyeŧy+Y>{oz(*L1j¤U(NBq]ͻlTɧU{`fgB)Op:97:k#,R^3^RRf$©a?\B<˾O>w?5X2i\;~~#.r2R9y~""fٖ偡ܜy ջ|u{G!,siX$S$=^ޛUzz4Ս]}acUӊmc jqhQo߬&Z>FoJqwpxM ήl'',?ttvYAA]~_mڟr4_:<HEa paxvus|jqpp1X/e<=2+Zh)B's;>~O$*Ϟ2޺ǟXޮϹ DU֨G"6 dMY9$+KA7r7;9~>5"_ *6wOgLNſM[/1?s<>pw;{_^{g [ݰJO,߻⮽vfU>[foz\u]63 2 ҂@9 Æ= 4<~)ϖM/y3L!#Y]%I 5*Qke}y]E%R0NE;mG鲌#7GA珗א>3aڋs$}(Ȣ`=4uvw|Z7pMtL"ȆT)^l+RVB^;\2 A # &ju_b2W-JH$aB"ĔS**ƉEF{CCƖVi4JqL.b6Q"UYF@5d9D 5KqDKZ N%34Jf@MVdl)Q_;J iA޽:֬%)LQV@V'()[[Ub'cftcSIx34fFZ+ HUYTZN1 Gx3`x/Z, uHJ$݁6l#M<8*Qޤ CdJ ]4OUVҚChȡfWὐ "Ƒ֑~BiY!U& 2X )DDц/ <$d'mu*H* A xNŬsވB %pUP J{'[lJR rc`DHX H`B6=VQ.%Ɇ0H%"!_&| EC*! r` %'DgB|@CshpA9]inDM4 _.¦0py Xa Ii> Ki) ^*P*Ri%e"d"@آy, R 6 N-`67[Ҁb=VM2]FPybg 8ȰI*z]C JKfH57]T?W(c)(`PnP, LhD挱Y,5%E L9)t0GRF@f5@^ 2!}W (I(HIi\B*&~j{ $h5y %d0Kէ }M2ёH 2JD A &FdY\$Don'49eToV%8M23#q| thWVLKQDf1 Dfip8! `߇Y;O2&哫eػ`\]ɈzD>Y4ڦB$a!ƫAy@*},<*&$]45-*d`ffSPxdh%_f`QB'sA-)BJ$RFV 2ʈ2O LBmxxL(̗Td2jnuf#fH܆m C`QF,TGW>^_EYQ"ܶ,9e58x0X ݏ~qv‹~wy2W-o1s%}< ]{ #B|u))` 5A^lAJPuj :DpjI7P@R@"pc 2=6A@15I(k@K^ /B>L%A5,|Ռ6"XXj59ഒZ25x(DxP:$<Ʈi,l$ fJ4Rd!P?A.jDQqEP~2$c;*gɈ&BP-ʉ6BQj%[?X#,YYDAy"P*YxfQڀJM)ISoET^ "HA mܭE $`F} %L0Cb9. %v#+ x\ﵷ:LW4 P7H7۬IFBnc k0ߑPp(rHuV5fsZ(ΣDj`QAI(' @ٻ޶W|J_ ` ;X`vA?-!i{4A{.dL,*/W;VWsik>s{8`Q"`B!u>f怇Ze@M%D.D*z;G ^GP)W1Y mC0+dH>z_aWU.9. ˃LTH#jJҸh 'p2˷?Y#9{P*D l` J .Qx"Bީq1cAR/,˨jqx<[{[()qJ*Vu^H%Vb[j*թN,XD+m) 6WdhWX8VZf0ak·g#t|p&Է= _\XٌCTQc07Η`oM||fG'8ׯù3sMk\8Ĺ&5qsMk\8Ĺ&5qsMk\8Ĺ&5qsMk\8Ĺ&5qsMk\8Ĺ&5qsMk\8Ĺ~k 1\3\kl{>k4¹~kcĹ~k!7\~/?n?i'RcrK@k3 N1!o_q,+M$cɧg0O#u>bFx5٫24vcYI9X))7P`ҩli$RP^`5/"y׊ޢF}#y[;] }ſ^]~=y`_d] H؟?&ےE-X=w|Ui %Ei4tB-{eaIϸR[XH&}O8VQSBmQp6Z> &ˤQwi|wUaDwG:l ެ+)ۻ^b2qrx͕k}B4_lrփC.>3"tiڛۻ3Z|@$10 n@zHT"l$/&0[Ӫf*EmvCl:=c\/lfBZ/ W3{}rb1[p{4nEN u?ЙJx^]p!U%M2!X%uMU)+XgtPVݴIT,QHnS*> ؎9_EyS8Yc7~ǣ1Ok7co=ym`wZb%`H.bF Uj9ϕs^ejdt?ؐx̭!3 AekraɈ"\281W,I ?l6e2悎db\>FD&oήŭ9 S& LO}U`>=4ϥoVȓ4ž=F*^o2X Ӵը_ Pte;r\ͼD;߿|wm%Zq.FL"iK12WN\0:D˔ HJ&Ovٓ9[-+W߲8M[$E[$E[$E[$E[$E[$E[$E[$E[$E[$E[$E[$E[$E[$E[$E[$E[$E[$E[$E[$E[$E[$E[$vŷC:'1 ͜3zb1I-8!n.fL܂+e ,z%:k\еc `h]ň9a9c<aOGN塀lWhq6E(`RVj1#R>6@> d UTE9bCf9P=`MOE% p&AH'?ȗa]N+nǢǏ|-шՐK9]k.6X,A1kϥw54= Kװgͽvex QgykͧXk/C_g}2DVScGBȕ1"ڢre,J-^&C*?% M+4(tN8mLM}ʐ}Jp*MŪW^J}(Bj\LnBo_e⊈`Re^U jMiSc:{c+oNZXz3s۹_,Ba^ fN)r GSԼ)rSwvya7NfV?EEeӿ;Ǯx+~U@[ (>~Cf%>=0wi!0ra.7\߰ƬՍ xx~Ŀ?LF7yxpC _]Rˤs]Đ>E]H_1QRVg~!pu [D/aWW}?nJ`gdSc, '0L\ܱxei7ۋrg˟0@>+1,?61bj00KѬ7 nޯ^_!" ^=߅>PW_[ث]FJ}9҂{]oz;#5}<Sw5_{p 2XOl\EĴ m)ΤS_yoOER1:ALXu6 'YM*8拞UY;kqOC+5 z"++s,/]#Ad,h| x[|I-Iͦ?M8I<:5>o$qvz+<2<~on@ ý@W*FRUHxeV;z$e2)fxxۏvgrq7f_d9F-♍Y죶 [750敥 5&P@]kuM 5&P@]kuM 5&P@]kuM 5&P@]kuM 5&P@]kuM 5&P@]NׄJjoi=V+x~QPpߩ<z&LgäT'l4y:s.M`So\W&oɃ}6\!}lnF{MNvZ㪋:HH`1H~W Ԟpc^y?U)S@̩ȍ-ndimkcY OEd欭WfOn=ЧDs"7*mW?E>7eC>6>ߩ=y09G䃸e}O7Glvۢs׆1wsQ+q<;{J/`zi(<2KpcXL'6p/nedXNe zft=!&ѰQxs>UOomѣpkyH+`ޏI"zz|7DŽ.>;x;6&BQ?L*x[AEWwpuxu7z8G=cnw?~@+>|~kWxc姻a 6u翺gCܺڳwI4f[tf{|^>on|בaWXBjθs~ p].w*jŒL eb{aVsBP?0:3X=M% Kg'KR *:k$|DS݂ٲJCOIkq ˰4kݲhl0(~SS07*bux* D3ײH,P OK:i6[d5ax+eް G]Ao!Vq*>&%pmd S1'iZ4\gͦs:;nzyv*||*'yr`{UEAXXp0[Q)[O"T-Kb G}~e>+P.k^W{{Ԛ)&i"pFVT+n'NzTFbBQ*%C(㒎vj۪W>\,9&㤚6`=>$L˪S%`U9bs\LZ5Ng82{x~T=' 6paS&y"ik!)QJ$5g Y"oQZKtDwVGMN)WUd·ŋN;onuV= ܁q.TRj,yk*fR*Մ5RjFIuW)+*J^,wzR]_ q6M5^p˷C*պ tQIET=3T"e%O;$x_ʉEkV+NBΔgѨE0!$҂[^+fĬpN9v܁9[L_;ꏧyLJV|`(kP%Ӛ`+u$ :sM=E%ϥVDXN Ϩ8<ʟWAԆEPpfo,$hֵ\ =cMy?U܁7N9M`2ĥ_2 /sYH)yV6$R-,-#T't™2܌SQwS!VIJ p5OϋٰڤˡWva[tʥgwJ:-m>={0:z.rD 7khB()QJ ǵ:h]э }9>Wȷy,p3],rV ^M|n? G5DN+:/n7O⺴(Ezo,G ~O,>50k{ qWbKnV[w UO0V6֟.J}_;n(>>ԠyDWQԚetuR:)vnEdi9^Hb.'r7T<:Xב ah&34 yFPHڅhSnn6LȦY`9$RD[nz368^8&GwEۮnQ}O扦McV^1NhPꤓ|g7n6&);p^Ǡִu] ,=lA?+|H&~жDרOg3Q$RbP5ڢ(uAZ0_u)ύc0B^ kQZ$\G Yb  eHTh۪-N$.}b"FrfRJq&P892;GC%Y@;޶68 :3{*)4AMޟobatnMExcr53T#-xd"qJYNsL2cP,*xe2&G~isɅ#zgt1=y^0+2:F@KyKD He!+xz+6ǪaU/0e,Q^狯J>ӄI1$V.,nl*LGqA n\7@\ Qvxpp؍w#.^AcDXDC9JdR,2uOtOy7j$K]ϣ?/)5͂â$\n_d?ſkڣǔ3ɞ15Œ] }¬-oVA|qpw;].S2=C­:\V__WE267d2_&zR}E6ZbSfiA%@zOns&}Q\TJQhrB\~Au$˸qD$,Ő"'`\H7YG裚j#8d_! >r)\0K9#1jObF$W$idP"zH+Чro龭ۼqmN6v&Nu8*m@yJsUyط ~}ןe]&Z#YtkZeF6YFɻd؂LSMz"SrXuyÅ[asf(Ն0q!\[jCʈִsWp (rp 0処 uX懡P20b[jשg=h{*]QO[CWm+DKy Q2b::FBkqZ`WO3NR.+W^1~+w~h'ZGx+޾yX$&ZqQR)*d*ӻ߷4*{6%r/&ve\0YM yUL>ckql\T B^2A3o dԥF+u DZ%t9e6mހ5׭2\)ۢ7dF5]o@No8B3MfFU5txetQc]]G;\05ti ]e^rd+6X\BWl:]!Jvut%Q:G2\h͙WF+u*ԝvut6B([m+(k:]!J`vutAFg42`*5Ѯ-gwgН #]@ }+5UL[*wg :z1t7z~B \Lɡj; ۡroR7L[jשWl&kWLC 6^{ֺ|p#x;BӈVi:ԦciJR~Ki x .p/_*Ϭ$”g<)7=&Ӿ('8q>BH࡜^I% p [̏wn5.5)AO!"hai)<1Δ <0~,+r-=\M JPjd%pJV+a`œ&YsM~'^,wI<4FƆTs6;lE lX[,V[4JWgEå$0h ]e5tjtBl%++Fr"ʀi ]!\EE[*kW]!] (m]eZsSDWNWeGWGIWR 4p5tJt(5Ji*UlHk 2BWUF;1ҕM ~r5uUi[*m~HzFٴ VD  *hU}WCW 1]m?0]mW>V(y` vz*  dOI ([}kW4A lȊ߅n7A pIۤ7d@1mb2J0pz#J &kUe@ Wښ!k}!~YNvB}l|y?#[D`N+~nKo8܅W%dy%OUe[{o\?G_.^-߫Qy}_]Qoy2^/d~i_hưnH޼ƒo<`qw{CY~z}8 ^-ߺ)׏rOٻ6U-rҾ4iHyHbocdxdwfGy$E-J& P4woonwf7s3dmcv/SeJZ1-NJl{r$?闣9hC6 kJ0iy,ze,aBpI%K`Mvn5bȏ+4KV1Ҡ7^0485-$N#iL%ԧGjFzbwYy'&'E^|+!G'|啓2O\rqC[Ȃtou{pxo[`v^O0XSow:_J^>ڝj5y:~>?Tjs@ npG! &85U!{m/fe`/2]M4toiN&R4! t e> \L oTcr$xB7I\r^nJ}hӔ) eQ1x:.K5Uսaұ)PJbWMtWZt{ŨiهZvF9LQռBi_b[םzߚv?/K6 .` WC! ^1"zEcx ihlʖt܅&lIL ʣ,? @BKlfӯ)@,O7^^["|ƿxb|IӢ6H8?㺉)xMuў`ؕ]ٮW49ok(~Z` dɵ <`tB29؞ۗsGcEKs5T/c/" ^D B2%-5tMJ;fQKb TX ):%y. >M!F \9wzq]"V֣)XD3rv41X@^\yfJh?]CZ ^f@U/I4N.CX| g. q[]lթFr<͓|9rݭyꢻg3. cs}1ZˏRYÊVP9ϺUNuzQ̆ӰC޴+Őݝ}L>˳_H ZZe@j Q!ȝ "D锨#,Jrk)"65:_q%ۏfPky(bfy{B5D8+W$׺-|m~-#q k58}@/C;Xc} 6+=3dmz?| nsRq8.mv=@?jR]~>`6*dJZ!(u@)~4o kHϕ %f1T4A4B再O@1Hjt@RD".q6 w@E\IQD jC6-E2t4z ״T/^UvQZbiJPesVH3+L$R悱)&|+1hx 556s%)DqR<(dzNU4K< dB(1"RD&IU8 4Xs`NrтR~I_} `XVAHTЊ8b^`#1 ,`@F%1RQ PpA"2h= __́cZXBt2Fԃ@Qז]m';.xӧC`CY8}*^sI`Q,b: M8<.PC<}DtNwCw+7zuw yqneM 9c4b4z64;/|,K^s&ߨ6o+iP^#3u~,/$~ _Q1C-Aƒ{J)ͳ!pIeBMpfeSiiLK/WEc~cp53d>M\8E%)F?P Mn|S|ꨴ)g^WQ(n% =?/r2?%9⮾5bbx{ 䘀?*;/~e%'m[{G^zwZ f1s?;mTnx(r/Pr:nQ?DZЄ+՜89߬.h}~"`6Xw63P%\X@@x87_(J:Pw}0 أp\laEKTWE3&Կo/)΄oirsq9bџAt4gqXO+ u b-Gp8ZpKeFV9ݘِOCsߕ JSEbSxO&h]!5^jS:"7 MA2\{LAuSMeh'SPT)?2^ƌ] @Qc֮V::vV[ƼpO`I­tᢊa52|`'&͈y3lbzlfÍnn]6+NLD^h_/tg(Ōd>JMN(Ρvp`S]/4JX )* )_W#//Wp| ,+¤\%]¡Z,4rg0ĠL)I2+:Qz}z[Vq:Kmxc][d:ąGlX[ҩHC yao/. 2A޵THMi&Z@KgҟGV~?i֟'JŽgt,SbrcN˻Vg=w낏5q!&eJ)C[ N!^g=$ϜYVړ D& vS˽gkwes$J(!t2 ޹2JI(02R/䬜9/aΟ 9u]FZ^N(넲v=}N gV7z"ePJ]u*|U2wq ?F(KRU):|@w^/#2(trS&XJtۜ<93I$ֵCnMpG#ثk#}\: wb_8vllCJ:#Z{/bݠn~|G`5e=0CH_J:=?^kݿ7ߊl;G)߾A>ctYQ]O?=L`Ϗ.b~lUAEO|&~di!&%*bdV] 쬇Yփ~}eߛY}oEA)C>h0Ea,wr6 zpYi9v@f7,1,` f2`~..xQ\\Lg@.0'3YϕI7sx}9-MD'N/ϯb~|>Cyy'Kaի"uRU0˳PTem.փGt0Vqvhǝ p;Zl2#D%`&yBp@EvZ Bv1@1X.^9AöQ6&C^sJU6(h4sۭC梉k1p5C* &ΆtW_يPw%Y$9LL)gOkj_LY$W~J{ǛӴ:k^˵åL:!7ڵdń`r BL\Ddb0,>('as\S,*ZX.A.mڍi{*26'lΑ ΃S3J(A X?p_jB5;,fuuUQDًQF1ZcQ%6 T3NdG~:ITSFu*{dS2sɨSm]bxP4!m٫ yjfV.M⊎ :ج`jMQʥoDW(B?!)DLqno'4ɴ7i9IK\M|g:WÞ:8-']6 :٥iB3LI|(ekjc5ˏg/8_pۍ9l!8S_NWEzo'?-;i9Z/,#I^˻r19_-ݓl[k NjgzRÁ<ϳeꬱE4bQa쐸-)&0M.m?A[=)b+7Ƿ= MiyiKגnJޙ{1`^Nq<7Ixq7?[%x8BHJJc vqQRIaeJ4/QN.zv2|Q赿~O{v۪ܥGӦ4W9'pT~O^t E!giE)_c:|7py%wSVHnz^pDa yR Fm}:j'-fK 5~KכȻhڲץRe&OqdP&,sFG{>E~mfn X_9pi3*ZH1 (f#2bldBSG@ `.;NbeB@U+d]L9ʬIF^k?P/Q |9&oԘh2ѼSmӉm6-4ۡ7-\;޶.#ZZ[Fp[-XK<Kٿ6TBS v^Yߡ.: J;_acACBxӭ Ѓ* |gW}\C*Xt񵘪 L5Z-1r +CyPqѕ2AfhdCU\"BäĹ&OU?ZDV[&J1x%¶3]oگ?5z{ؓGw'E6]#|&y^Q_\*%cN7(>tCj/_!#yxAVEPiˑB}|"p+zq؎8cSt`98S>-Qc?fiCPy%䝢M]G_7ҝOD jxkN컇y2g9lFYr]!܂ıpԪDk"3 j,T( Ѳ ]tq4!MעsG~fQG'뜗kh).hYB>&0KW@Pk8.8inw*k_)BL% }*L Z(Ju\F1)־t+a[V&ӴD^kUJ G'ZUa=&c)l)6\FDdk|E^tWo0Ʒ(:^lhu^^wp5#T.*PJ6y gK0 1lkn`p1 z9=#t)˝<_n$Zܢ a{8q:^5| jg]X4 < V%)C%)8܎BU:9#* AO$vF```T$%9H,3 бqܘK^$yQyPq`܏SMENӧINΛU%,)Ą蓒bU ( ~KoͰdݱ.Gɀs߱m-V-orCLʼQ ԛ <) $"&5։)kl"ZH*T{WCh{sA]LtzJ)]m\T%Z141I枒t-Em m$K,EWfmll켤h>"簵i˫[ֲɠ1L@+% k-9uD.@qCTC 0d] (϶ŋzO\NrL$̖? b~cΧGeQ<>fy5!q. GYM8wbSkNJd< |5qOۇ)~MϘ.C!cS8AǶQ7y j>wpT'C$¬3I\ڙ$-ǞDט)0ߥs3}ƅìW؟^;Mj 'fvL.nnŲ/Ÿ$u)vq+bH!CPZFP4g>ar0IFW__W✾?0/x: @IA]ߗ ڄS9p;;V{8kA#y&FTxq 'B`PTXEA=G_HnBY/&O7X*gޜ1 :STɬU*ا [U˻q/bIf2Gшu|~-/[p W|7ޏJV :-Р/m\>@``oCR5`t|}ҏl̪OɨҴwן]m\)u \?MQ=LFla6 >KtTlҟJ7*pB]Evc \<3 #7.n9*`Ӂ^ˀ1T)W逰 ߜ *? &7qq7@nÏrVh;_S߿[U|hZ(_7ƑiO J9|"`8]W^||&@%ݟWbئ6:hfy|%{0F77MSCm}reG~wk*Wgaֻ ==ݙMQD6Raﱼ$=R$١P@egBI\HRq졎$e)P+ u[T<7veSܾ/ۨH])<9;#n-qu 1| P`Xs0fP'N ⧏-,^x ,,QC4aq8ޥ=uVJϿZ=<]mIh{)Y#elK9#@]FJ81QcYD aE[1ar"ȼZ%[d~2Va S띱Vc&ye4zl5`BZ"2goS< hW].uU׮ɒOV)c G'm+̙skrVqqg=u}B Xk&a%1\KnBP31k@)IB AZX,`D79N`.0 DcXXHjk^C XaRXp'>zEϢ'+/d%qvp p}~JM"A'g9 wmmr~nGK);.T{sc? ߾(*S,8O QPvbQ\QAYH#6"P2zEi#')&=)Q[M*FfdR1g&H*aak/PXXI6IbH\UNn6n T Gl CXY6A㨌JvBdNi1Q]dLq nd&TQx 3 j-$& fh/#vHHLBDN5qv# }<nM;Em2jj/9R2U !d17yX. N iU3b^A ('8Fc}9p4e$`<DlK?DDEy{'DS߅ƑɅy)Fտz\w]WTKjt@J+D'( 3Lz\Y(Ej5)`8B ,MXyh{`0U\K};lyDECdK/ڝ3Nm^HΘdS#1|}V 쥑@YpHcL!)C:ʁHM)AV^e)AG^#vb\k4 qThz,Wf$}WI/ mo~,NjZHs$,}zf˴hnTM6xϬ\8! [h@5 )bzg 6ރ)tnpfŋ oU6)X(fJ3KXB$d < ឺBwa8 ^AǴXK/&ό%[eya3!u")+ TԉOd>ij BU(.4Fykk<#9 $m#GEiE* a;na'56Kq&bŲ#َݲ8b=U,>ٲ1QU=p"a`vrlJWtde f6c0B)*t5ThvosYˑE@I(ލ} /Y9ŤoMwzv7^}_,BUaK+O_P)@14$LSd)g){ά?^t#&  ?L@`;Z5{ETպo"'Zs,u| omVek6~"Zdd.':9}'-6-iO Ky0~+"mbzG@=AQR}0(fsƵn&,*Q}Ca2v(Q西qEmSbONIÍu_GEGh^T\D Tg8S_^< }_m/*dC|e˸p)ۘܯ_bt@[_S'Jp/SH͎.lވvSYq|*Xveu};/h{Dn{MV΁Qe)0;b6/*JP.F&YKFb#p{ihl^{o>uaLy1 l.֠K G‘OKt@9Had#8_wޣ5V^ 8jr6FP "hc0$ádUZpVg-&D`"JUGː֡kx 51ݭ3l7Ѭ+i,HC tc1br{.W|yxN);ŜDns4Nj^CxFxy28 60kB¢)$ \nO9x[sK,-^kﳁख़}~|Ñ ޡAj7]fG2 IaNE';iϿ^t@6 H+7)Jt&H;`* @4;9P!QTE #Wꨵ'yTC#Ze׼P^qQF 1tZi~>ItPkŌAb*9ArT 8z(QHQ1Tx5-'ŏ AQ &? uncMŔҤUDZ[RX\PkUocɫq*vZSp:YїPՎjfօ4'8h?M1R*ilމ ˋW>:Ou3}10J9rPq6`Vo *Vњ/nF!cqS܀XB!]U~i^[<0Z[ip1rVZ,aM$ Yɪ0;#ٽ?}{dJ7thl~:2ut)'q˕!9cz72+mq#Si;Ge D6Zm5,+phԼn9-`6esLnKr{ˤȍ&_Ot2v e=ޣ9Ln[rw(^cizuEjcɶK/5K|+rkZ O۝ͺ&jN^,>j`0Aj8kpm@~m&>xٲC:M] I0^VY}R2\*J#CrSTrS'<y(9y($Zr)y3F OؤBT'*!㰌ZB?-\MjWe5u:X_ PV.) %V`yU !?cw],rOVQmOɨe~hQ>%CćwhjA5sA|8J,I,8Pk6 ˟F]a R^@1 \cX)i%Z,:A&54'`yX⬟@..uמڛկ](ؾݧXTpW+yB_{,N?#ZU;I#mĭя|~??ҟn(@={![kC.(1qf[L p _iIdhzRu4{tHVq~cUU=OgGٽ%,~o޷6}+tVsq\Vl-DSO5ЎoY&$TMJqEˊdj[ ûZDTP3XOSUĠdqtjq:Y6rT!kX5˭7E9v]BHCv 6GP` QWz=$NQcnGE[;4m="-opϤ;3ڱgv8TBd'{ B)av>$ȖDJm@ŪSUg:Kׇ3 ETZ"s)/IYKtZ e=+YOy!GPe1%D#GT , ^O8w˩`A uӕ<[rKx~y@35΂sz(HC< @ l4*b Y HE l߿D%^W-VGCR^JQӖ] 4L җ`≀Fy1pq<c``*RGiɣAA*8b Xؾ* "ueg$ -!C<HP8AGxkIDA1 RHI"fI"8a\?}5ڹ4ß?Ywعau]vX6o3òR`*{.y;i~Ib_dlNÇ'Aiuno>d]Q5d";L>{:ƨSGmbTi5l)qfn/2x̊G?Sw`Sc@>JD *yPO|_Ɲ?|6VλdN-7\/kUoF*)yg&|rC~kl[0鮭V}:>q7 ~ chnPfn8c+MN9xp A2zMA8T)Wi#1 r Oaߗ,L~.Um|f[vO+PzÐq}_]ūY4g?)=(=LK\g?%b;%*QUvQw #߀uw}/CKk@I_5؝r9LuǥWa1 zr%BߺnP7K^քޅsϩƩ]K^֯5f’fا# &klp1U1~]R&y MEW5&c~RޡYmnSX>0H sO3X= sGgoFuQ"=sf]Xvqoi8d O@4qўl{RTzI\{͕ 8qsq\TuiF2/J0"ǐrZSII#~y<{j0N^,.P^3 lM\iMr(xyV8.$NHkZKyd/yÍmJ9o WsX|P*9ߊpm#(!cE4g! x|| D9M93Y<9);3F9?L3ELE9obm,G0AU^0)3@C4HL$${zc$(i0jcy,Ww]Xt<^2;{iy)fpx] BNc0D0`xJ{lSn՛q=b~AUةv}"U#? D \QV}@5FM"o 'gW"~~24~P=+&dˮz(oJx=@8ͮ[95*祵?doRfZaa DD/ͫlTJgV0 1@I?f\sT 'n(*=+L3 06AM+9uɬEpsO;i2?dOKum{9G^Ǽ~5/`@gib0L ȊK2ڼʊ>whSP[k5zk(RI9𘣨^P $En ߃?[:=cԊ -c[9cbN!d< F$Ai vFjek6xLU (W$I|&y¨U1kQRk 0L.z=멙b[h& tf]njU 'PC5Y`@5 ).ց3grǻr)(p[Ubh:rHitL*"Q*̀I(͸J  B9QdԽX"Tʌ^.ƾVʧ&֍J&K"ķ(:(m-fBENS.!- TK5uE'AdDfIQ\h1JkA]^ŔFn5 _qm-!5KH[ +ѯb5&Fx)82(,"@B_}?t⫡F1A娯2pU^o/'uPҴ "몛c~z0²T&6B e}yi<+L=vG@U4N=_ꀽvP=<)0 bR~q,8TER΅`E|'_-@8*'nMdkz ;|Z_4t99/R/[ MI#; F"ryEf2H%\S/[K3Tc*F~G§P*QfK R+Nf'pU?a^v Ox_21 g>NtQX3Snk/ ?h_byFy @Uj-*~5)0 (S=.3d iǶM^R_hyqj ˜PduނcL|Mؔ ʁLU=hny7$OL>ނmNq6ZƤS%ClM:3ytf$CHSK7d_| Bls(#"]ܷixOlzɎVn g3k>4rdH9T 41,G+O605!ɠ0ߦm$M-^lKSs]v_rL3*("@1@ka I@X'4@)-6BhwNٔP+2J/6y|/gjYFm'o0%ERJ&p';AJ&(۸()S~qRʏtQKGnL?舊 K9K.so1k]]lc Y{^>cJ s%Fy$6dEx6am MI`NaN8F^Q1G E*K#"( -J)nؤm MZ5O~zWEtylCzmFM+ZOd3Ay*OܒxaB0wea#ƹP+ƒЕsD^@Rޫ)†P,qm9%X:x +-ܟ6RXL x$ )[)cƌynhDk45Ȗ۱AGW ~7ݾ+z2d:( :-rJw-}&%wt]Kͻ ej+~1kA+s(ISUq `I[iT:Sj{X~qG6wDw,;9klݼu{y;7yz| k>]?&8?懺O؛pYŅ/ށ>L1^0CM7 Xq[C2-6"\jy9POdef[|,g N:zfZ ;U*,^hϼJhAE y Sf(|`73URPxXHUJ'SݳhS# j4!1,3BRϙn3:} Tق|ưf!Fvڃ vP)/wBRXh#QU"%QlT"룥,,8Ý(7fuN"NEWܤbP:)[nO&M:?@&d$8! Ҭ:<}G))V NFG M7 @{SR:KooJ QJ ))/NR#i {'4H}g>'Y&p$hMIIHiM>05Ɯ-JC@ 9-UrW9lA *BYt ^+d-:4:c%tC4I2LbF_eEGrS=sTf#tT;bv6Tx%:YӒ~.f-1˓ M2WU͂[ޙ;3?:NU|M`/F`46@Lm.n U#J-lYiCD\A"*c+]uuyc}9%}6y+C&'8 v{풧ߵBۺuJNdNw{cy?3+@(E>>dyj%lz:F@ҥaw4eCtƷtVIg}XD:,"d IȚ1S9=b4pMX:S(}^w)z2¢t(Ief >y .F(* ҺH4X)=$o"Lޅ :!$5 am"easFd˱`L KXS޼86J J{^\Td/#cjSˊKHTCYY>4?1bÿNU.v65l}grl|0$g!3}`@9 VmOw>:aݯ_~,4,gL!@dCJI&蛢>$6mʯ٦[CCmZ㍫Chp5\.튳[vürr΍t)z +-Yp#U,EaKg:]4PtFBZGKފamP ݇cw4INQn|A iE۫_ƓQi{wλn=;Ы+mZu'mO۾&^?/}W ^on] z G6\Wm3lzz5{19kMڝWه0^)AO\S>E"֠6} ꣯A))[ Zd5 T&`bq);T22F@Br Ie62j*v u4%L-lM * I&!$_l8;g1f:1%}v'r5ϧN}OQp8_٨Gvqbo)VQKfA'هsFrV mJ֚OJ^1om|-E7)`%BA:aXJ]sv65^ .6ɮXz#)%gy RR/lz1,<\gKazRK׳TE ,v\[.)8 % *QjlԀ@ՠҺ"ljMza/wext?7iW%==dMkL>{ B PJ\|M@,6AL}oavQ:kAc4#vmqM}~oӗrJ]/lʓŔ:f!槒>HYoGD췈RBE؋E0Ht9!ہbQޔXK䰨lc3.H:G>nϲIE۔38*"ˬDVoVS+kbEF.V% kގΆY̱[͖ʕ9#tCYnKWno_mV7x`R-_6Xli YeEUTϕhNEr!%k*Jy tcx&yMV[fyT0 Q @kIg#QD^I:htrAY['Z CU?kB:7Ak˂FQ. !k@ce!kbUg(J I,@JXY Wj2b8:bN;};XT\-~᏿;m Ou ޱ9fUL~b#4fhk~۬1]Ջ6ɃMhul8ߞ3&ͻGn?)j]{cAek})3ϔW~F߷O+;6]Kcק,˯'b?atAUA nILw{>} N24ۀ+l[lȁ,p oZ"v7{y[_w} 3nfjM.F'&%׳4M\[|,Ӎ!wmqm[O8bVᑎu.Q+b W*~/NןԞ_ d5P4̋9 6 zLUF-VU˞?Ue<ޕ"=q-]vgj*pQ tf8GFz UgTFlU^}> !9(lJ7gTOJl ig;-iMvnU?H]f:)Co8./4xOdM6d$ '|ck̼ӭ))\23 Ѹ]7Z5vZF5o38 m;;;u|語7}#ˍmI}>K&Bu-ņXc7DM^]ʛ6lzf^`o1WMui]m]کˇlk;>kI1Ɠ{m?ƙ/}̫iզ`47󘎠*"=0U]޹xHokxp|Ъជ^W5!jx_ z3nwSUՆ͗bq_:»aNoD>Ԁɭ=2<;yxمTϪ0Ix}eSt qs[L" `M"5 zY&AzS\ɉlJN^TΚ9-.Bߙl<&دJjq۪cf3q*|?a sh(q6Z,p~Z%s# Ct#t(A]zRjhqܤƓ؞-4I&C)|WXV.٘1'er)}hVe]R l:*,XFH5En3$D yva{RWRn)K4ɐW[它v$cQ$ZrS 0e ˟R>[Ym!\&"ƾWֶ#g]Pq*m,^J{Hu HVM@*3Z4?gR8JG h]oriJN 8U -H׿zŃ;Wzk-b׃;+{40еH0HK-9-{1'?wm#I_!..Qm v7;X.AĈ-y$9&%YLX2cTHVUu=\lWyw.]5.:$XCIJtdAd` {_i .mU-vh ec)\ qZJ~h*d[^pD _GJ@ks9ʱ S`# #3$Ӗ3f@p]A-&$=Mͅ"sNz̄'jglmzԔg*nT:D_%=3=5=@$=`06ˡd֧NNi!7f67#5 j[y|Pb:Eͱf^J}pA ْqB`xɍ2:&Q,CRht1{όr4 zR0(<MP8%]&;g32vU:4cW,4c>bK̋OLn 3^Vl g/V;xoOxGl ǹ`5;Z8Lߙ܌|wjuܴ}ӪLWѹZPVn,Vftj4BZϜHVcx³móSxNt|` Ws0cb ^g#eN("C OZ6)CxH#`d1BEg}N*A'ZK9bg?>7Ɠ뙉p?m+8{ 'ux2VyR~k=_nh,T9ĸ#)p3+yͤp٧ .x9:\o3makKo(\Gh&L&҄14;&/CM*/Cm"o@v7ᮯQJ\`ઈkIt*RZvWʖg~.z}bucc >W-ɆIwH}a`ԹLTo߽FI{jI9MjsN EdpdwO>0Y0^cTKfxm }t(/茟垠皸fI +f] ypIĚ6 *BXtjK6 &Bo#]cـvR7Q=[|V~4QPȍ?0^ 9zsٸ[jfÿ}|hXmndغ}E/kh;SAl*K42UTnq|˙^Kл(ڲSeZΫc֙ϩ&\6OlUƗuG-iD`娘}čQP$a*QV 0 Ipapw$d#q{n hFlfG-9P!zZ5bwʕMSs0ԇ]4Mtbe[܍S}]. x˵ƃ.6Dzkj@fB@HWK_O/_N`J6 M-^pW#0dU< \=u4) b =)w} gEW(V9LN"9QBIҦ΃SZ1J29Vn,ƂTreA{P2<; .iRglIav]}Y$@=k`:ޮ,mέdEG'_=w1zkVV|.e i4JJ]Qeieg@ hE%!@'o'w~ swΦw8Ζ)0fJ{f)W7p =7/dpܾq۞x/ +1MQ+Y::sΎ'V̞7Qkӣpнh#QYuܙ4j LiPmS B[ e0֭\B(u)Y[%`T1&dDf[;戝#BOl mecgθ+nt;vÎZL'ks|) QL)⾚L{>\ug2Fn| 70:h vH/_!#=, <0 9Z\Ph gѭX uSzSr)0 ؔ [$2BEwV8Hdϱczٙ8[M[F`2mW .j1p]O/ 'kp<{k+\o1༒ŊVMg]،YHt0t~G:=kEMH} Z?t/V5]7 :?gC_4̚άںNؐXH tZ[ɲFZFeƻM^4ZK >*}u$xBR'i4თI:+q6{ )5tgKEN+rd^{0U.|T,7[c-lJGy?ػFWª"%$\^u!mcH6fjUDl  z 5 1VcLN6ފ֋`}ы}1TG6mPlLɄeAt";'i`[*K= >4b&Xh\H#[G굛Zayu8(=>* a{8EUwF{ʰp#D)'#/ LCB&9)$e݆0 (Hdl5bh(wV+YfR8ㆤT#ԲrjOJYa"(0cF4V} ّqhh7qbL|~X%Sˠt[l[S|h4>LyoÞ%˧bJp9E,jifX ]zmRS>DQ@ӣ(UL1/ż\{}|>6y%6ŶP%5&'LVB87-9eab!Z3ni(RX|XnR&XP 5PS+cWM5}YƢJ}bvtrqEcɄCb6Z\Xva,{|7.d):9 YXS59FDG5FPNKubP >,8u4ٷ $r3%ǖgZCg;jnntqѿ*wqߛt)=N >as[P.|_ /Z}5Ҁ" ڗx4¶B lA TDU+K#ڔnpIO.͖6V=fy5 8z[@9G{Pxߞv|ghm:Ye6GnfwgzVa1QzސkEmWD|b1ӔC[~XlveWH\0) }]j01r1L0ctcIi'c^RأM=\iz%jk>-[Q))-F0A`Kx3Eit\`rH"Fg˽pv>})~܇g'k!~>t*Uې]ι>5/O;5Q * (M{$jrR0S z`mi* KȤʋ=dd 5c&)l`u(mJu:YJ&hns7lO4lS:o|f0痳֞wiw(/#H1nղHt:RL,-Go)L<ȴ}z_~nw 7X(rל2 Ndn@H. Nyy9)~fs+8=)cV)W=ΠϾ,U[2}_N^tppđ$8"+* Ǭ_o~Ϧr֋Cc##t]޳֓cT!XL&8 Rħ ŵd_!R\Д&6o,57&kkXL͙^Ց^*%w1dvWl*\U$uE) J{J/OG} Awӭ}DztX_]nӲO/>x#F ,.:=sKbjM=E_h~Q/z6|4r2)gcd xP?6rϮ_2CI'Lz]Je[=ǻԌ q) sm)yVSW(%{~<3 7#|n߿dKc -6c SB(ٝ !GKc8n~B y܍0XuJ}ulZJAab6+RmP%ɍN|D􈓴yrͥ_Tgr>z6QO{p4?<+/%_ޣVlgv'%-} 7W?z'BTOG=|Ko:z"oó/GrK{NcqUzEsK^o8*xȒ25~F{:*G~}ځ[e>4|>9j;Ǔsރ^Pp:Y}كu`뤟!Um.ATD- Z0ڛ֠FMoӐ;d4$|\6#xLn)O=4s2>];1BG#II'a.r`}WWs(2 m1?{_EhIDJ# ֟>g6;XX=(k#)pic=gwwzu'X:;>i3`zNx# bR\,4ȅ(!~Zw]^EchL+6]Y?SNxIÛz!0!x%% .6\2ɲ RR68n½F*MQ)Njil-Oxr4.T1FPԭ%' $j)H*&#^RSɓu r4,پAM7S^Ӌ ˆp}>Lse}֓ 7ʐ1B| pvg<$P:\Yӎ@M `|//SniG5'g2~qNh 11o yȽdqz߼7}PxSrrFrL97?wwz}ry~^oծUR {Boc(B!e/$]8ގTӴ \@+uN-z_5G5O]kS-w}# BIG#,F]\ѨKvQ/hԥD7u4P/u4j'h3>\ 6!DI * g2r΃Rq   OE ?HC.O6F#{h^ĶfY(VޭBԢ{l((lr25/za{zZl}?ݻAJ'F 0bBxHvyo\G!{_M6U&CU_#ׄ&m j7{ۓ턴ihvWCnY>.SV{Um֌SrQ08GjyjI[j8ط0'8YfXX0D\^2cD(*a=N>7Q@QJsfay邖{}j_Zͥ`ĴFIfL>+8r\hlާ\1=^_j*#-$&-^DˁRȦE|9XjzV(Y2Rk!I bZPU^8j+CQ#Is`c d]]sr+Tbz@*UC*y}JVVt%y}T{HmQ4cn%C}4hu08uUiqYZxՏ2LɱHv3\}Y>&JeJɛu;ro!7-wZ]ߚev??ɑ cfѢaA9@$zCUZVTP\5 SkAnhHzj[WZHZ|k]-Bb"@RC3`p3pZe0.2B?0 ҦVͧw^\}[g^xvv'glPPIb |(Sg$Km1{[|յD:(cop\{V&%mtM#. 't(!S>4}+\؃lcaXvY{VZڥdt0eY)!dJcQiD2\2vVXokV\[bdPȢ\SʢY Qx!lQIuu1̇χˤvWR'0Pq$#j5,#[ljfˆʹXLvŜ}2 ٚiVCԒw+Ychà<#7 &Lq[TR\ 9S}ʨٙ舳Nfg*:o댎is0+9H3/μxSJu^f*)b* %U*\`^ND+zŧű`q,LJO@a{g /cI<pq2,W?l_ %O~YMJW1E΋`^pqqJH*;>z0OƱ{Ga~cNYU GeZ7tk%6IX(\bb3죧^k?Q|*B6Li%T8|U^"G VrĦ -ꗼWVm[qK?q!Z[a|e.5zojζՠ v܎mINjt@@l/.8-:Kg5O%?[7MJ'ߕ}PPzrʟHƶյ-ݭ!;_O_?.}3S2Pɵ\1A)hI` 6&Ėjlre;I&n#ϙljKyO) 0@*{3n c/Q4}>UV<<]zY:m 2tuJ+㢫p]zHľ-I=KlbmʋeqɟRIZU0 =x;k)s ?fuiAoB>gw Ibӏiwothѱt+2Zh4Qi5Ҵ7=nNϖ~rKfz#6 X8>cؽxcvv .Ymz[iRtmbnmX/ cuW|(B ]N]9ϝq͛O//Yc'ncKUjcbd(V׺YT"M۾C'WF<ӕn!$}zIM/LW嗭5p Zyp#[MtuГd UT h Ur:%LW]<)eO(c o oعv{V['Cx0M9yvCx-*2d8l H̃Z2Sxt&WJ&Ѳ{&QnD3דoR+'CWLf⩣ jtF^!]4O ]=;--_ƛa}o(o=l׳FWh/Z]˫e-xP+ǫIg5},z_"ǛFOy>ߖhU<[\\\X6:1A1˦m.1B?vWˎ~w؋57_qnn{?sk\ړ qZs)ގ+o||MwD'͛3/@$_Q1F#ep@Y >xD?/5?g3@#G4t#ϡ)~>x]E\o t!j+RQo>GcxSI ,FJeM|$ -] r؇][whkW|7Z[EHY@MJ[rNy#$Y&b*6x+zk/bbը#t!lEQjbJEbt37|aiL7#vPcW&`,bM%r,ג-OJ{!ѺZcPOBdR@̕4QkdZR &%-;Ѣk 𡆜/>~tI˖Z6fr\tɚ%e+RS[cUi{,B g06`Иf n$(:&ΜUWaE{}vM`&$BmІmθL֤ C)(xCShx5]ˡNfYauAV"vhH~7g1o1֠- 9uH )Z)qR ~:'evNԽ*S55]b9k <*BIT0ڪUɵ@)jE;zo5?$@wc`;b5} 6GA?盈iY]hRK]CU2X !$ZbA_2,$dN1JGZRq%;ŸH xVEO}Z5%UB-6JȤ10fr HI,;*Ht=KFш +!oTD!n0Le*:˄wh\ fXoY).!(^3V4TTPtT-Ή֊}T"# /5h;1uVsP jJ#Z,TlJһ7 Š˓±D6|)$]`K:(tŅexCl(4XǺg*){R˦ ApowU!( khXJKI3, + ѻZ8Z*c:]` ݐ` fQ\5,%Vu}ϳʒ&T D7t'X,xDu0t|GPs/%މ9 0+CpȤ`Pg=qY  rLh$/YzR AEgdD]vEVSpDFTE>c8 8h1FPTP]zqIdF}Àm:/,٧zӚxyw\p |gM˼ LМ'hX G1^T^ʱ PÆtd鳒ћA(h@h2@n]<^1k/,Ra) >@HI&bZaD^eHa4 .}"yL\t辤#$h5w<om m*^,:7PZU`;*kVd%gF^V3Hlwxn3vy\UΧL,Sе2kшwPb vwȸgz_}.좧b/=i xX"y,9w%Y˱JpbuH~<<_ <IO@ &ow!}Y]P!(["DHʠv[ ,#m6d B;1#,08A9DFgH!t9Bhs^uc-AmF΃@w:DM B{F tz% g$ ?AjD쐻#1Ȉ"#+Y{"W\- cD8gQa A2֥gZI; .i-,Uf5*)8Y UxV#Uu]zˬGae,WzMw $afs` "}AWN:K `. \S@`n y nu:贘4XxMLgeùf&AGd PMBV  Ihd*"ltLPPv`)jLuZ4kHhS@r86Xh4v Ƹxݽ=*3|{80(C^b,޴Rcl|(A#ds_k$V:(1˥P4LzB! B*|,Az|PPʭ`j2+  W&U.=7.Vr8/pN / ֐@ DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@:^'!9 Z _ Q=8+r+$"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r{c@[m5E݃w:B'0CN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'8>Z}ŭV9zn5ח zRb~旳M ub0ƥ W*ZkݸTQz2.q铮G̐޸VNCWCC+]!]yᄴ+*\=[xE@Jt5tzq$_} 3~p|^CWt%\j& 7+}aBW@k;t0ƈ [o^Zzb ?G7:A1;ҭj8YW:D#Ԉ鱖cd5~FЌ֜s膋5nʼnNxo~|m dUu>ZY;e~jYkQ@Ǽg87+o/ūey1I&Ph{ ڟ??UK7IZ\Rhfr~k9A^Zt<9/ugk\Q)Ƶ7`3Q xJ1P*?PZy躡t3b@t%nBWNW%:FRJ+%DWMd5|0weJstUQ:Etut|H`쇣򌫡UE>(5y#]- c W&h-;t($:BDWl ]U *Z2wut崳fH;}-:}U VwU]!]y7NO+l8*\>8x(%ݻ~J> Cf߼`/3~p%U=d?[/={Е$zlsS%_dGB yּ>q~}miR+\BӼ>O:MWM#M'l/-\56kJIi߼oOu+'lXNҍG#A1}^'UF)Y Sit}F}jvzY #kГAO\ǐn/j WmM7T"VK>zL0p[H)c7mnCP>gԳ_gw}1^hdV&9'.0h6-h*\9O9tt3i%|@tU?o/UE+UE1ҕ[*`7e=l0+Z*J#EWN2X~'*f*J`p ]誢ܺAtuPʠڃcFF\Wg/; {I#ὠE=pwWkiae:?  V>׋B#p{V-Ո ͕*ՁMaή ~߯^B-'R|Q;_!Jga??i^ŸCן-Von:}rX`WW/n>|[^v ~ ˳M!DMMnU:jUٿͦ./Ob}b8-̻pq r( jbqz;uRTBJ͹sQZͿ޿s8( ܶ%&T[(KHieJ6L&G}L2p>Z|| ͥVGb{YŃ柏_(zEcE3dN0]<.J27fxK<2u5j쑙,N6S\K.b<킂Ζ _>VS/?r-ޕI{}p<o4:CI6UYA/]MvY]9Z='hMA~v,X? ؖǨ OxrbG PbsGbP^X*|>T(,Q!Hc9S1;|aI֙u6{KdEEY0AQ\v2,0`{y;7;+[#XFy%nFxr6[K{~7g_uwOl?&_XM@u-:jxcJIR=v $h=YG=[ug m^He}Qev$>]_DdKJ뙔|kqW|6i6Y?vш~xӃrQ6lf34!He}exn W:Ux+|? D--.&J̙$Z.\`&Zq.(Z譺&I %ӊVew[l=QRB)FwRrT\ƜB+UhY޶\6/ԟQ7vV\H<1H&e.Zn0ǘBR|)@0{&6%pI^L`*663T<ʌ66FV_[$QPk[lɆ8 78_fJo\[d< ]\h *nպ-/2>zWmbW.v7&Lb36((3wCuYdiUr$URU(9UI\3,QHQiSFe+|-r;|+/ʛ6%zf||RCaޢ㱬{fmOM ވZ,bhdUs&do)#yaj[ml7\7?&(Qp8oIp%Q3n˸ݽݞ)nYͪ"(!+2T]D5UL Vl*A(}smZ"p4q9*WW'Uc<8~#Z3y#>䂥WWr3Q{R䌲5לj}jγ4GjnGBW#L$ TTJGһxGgGgҸ)9~&84 P}L6'`|-\VȾ7JUR0M-?RvC7?tgpa,565&/?-"^M~#{2G=~fw$) X8_Jtx2SLt<6BiS%-ΌeC.%WʆPS3;a;;|ǽ3;%%#V`g>VQjX Rrl ycj(B@75T!xNr_@ u2劌^Iyk*@%W~ ݺۛqa|B/0 kvL= o'kz0;ĀW_ RFǫ+CFaQɇt4]Z6F6t-lpn_K<#,q$H73tj er5W1TkxF  Dۨ(ӄ\O,֚Pr9%TkpciGgS5koz(`ƯUV.ItI;!p揸[n(U3%yzxc3#Eg=Ni8fW3>{6Qi9%) m[.t]_Y?roAW|}quC\ 6FZl[Ld%BY)Da5 ?"B?ksHWeQ}^d56M+.Ԑ- %gS(X1EGG&}]饖 Z %M>Wf9p-V"h̹LH%O,&Cא>^(T'YL`9T +4(ZƇ*Hzx#zVI>}Y/._pTj'(>p?99Кh&m|2JwS[F2}Pw?K.6_Zz|ݱZW8^xctH]>=.G=|?ov.zKo>R@ ~"=73nfJzX0}wi}9pfrO]"wNNiXQƬ,X$O~ŭF_kr~=W o\Ոߺw}B11]u[c߿LOg]ӷZ2/~+h;,%{-@R\}OVq|tDž=O=[-=Ջ^v!:KU^WJ~cԳUN{nfutsٱCEז[{WU5}Rg9ZFwlSѼeΉFAGA <2D\"G_4E!VuTUKh2 4Zan="?0[6kU sYWM3ce9S{k!]Sb{gr`Ndw#f̷W}͇Fp˵1E4ÔK*N27(XL..[iܥt荂DSQ0>[K;k_m<:<$Ҁe,SSB{emAOJz򕍟;o8((fTK[-'$a#*g`:\(bf*A%)Bll 1GQmc.vdV78YīJJ07UɓnY_TbՆ >{#ڨ[FNGRW[B-A3'+GPfs'xdG]}˫7T$=dK8~r#dڅӇ|;c^)c{M=-rP2lBh/@)lQl]w,qcD}Hh y}zlc/1!|O]1AW 3l1z)עzdZ$GL'a/980A&lYjDVO8p܎8SSu`9z?_e 5 xnz >x*o|TY.4Os.ǧ~Nw*~1FR1AGϯ^olˤ1[N|B9U5ݧ}~7/}vV>sͻNmr-WŹۛ SU?r4a_ Ć2^O?s6qWy ϖ޶:ns|ryMd6ޗ9=F5zBa)!siS(&#M|)Im饒K'KjTc5YjUe"y^?f"* FSZ4yq 3ݳKz‡jh 8aH9yx+AbtC%23:RS.H|Ȕ㐥$0K~*`|W>P>LLkk)WIR˅R&'.FB0ODof>S&Xz*DD{G'E1dm?d~;&v}!UEH[APYD=rCVw_THrzN &F b ڲ. -fK4Yh;(:)ς';;Cl*9{ԾNXIèvƦ@=aTW}:ВM+q|/Ԓ:&7&Ύvvxi%KQT}P nAδFrpTۉ\FHUB2%S dcEz?ߨb*!Ƥ755ʳQ)T4$YYwSee0jӉ'M<^Cڗ-Vϊ̊1l A0sʆ h!6τTR89QAUM'~}!m2/EM4LIXL1^ҵf\~/!R( Ɣә|mBgb|ȾF6U.(6:&3%ŲR1֒>n"U<.J jdS:FӐ}#lg ~]- oΐ!|:5;=d0%R y <to 7>RJJM\u@2e hQ/M))7@`T +-26穀aW@th+{ݱg_Kg($"AJՀUml窳RtS&7R5#;ZYeRAU *E͆4XVY89TM(&1GgMeP*ozt|r mi}X=?op\zu]_j-Pc& V}f}r"ŻdA}5'Kn:J'5ZnVXvEƥ7u_ꁚO]ېw9˜[&Jy%̖xcLƱ\:`qmVqv!MFYjg~ɑ;t\{E8rOšx$>O,۲gQK-5>bK,2FVg !Q#xV ƊLXZml\%ubiS(N<VP7m`OPJWZNV]֒LZk;MZ *"]_:jl]*ۦk<2XR NBIMCtеE]WZ*a>NA9[uihWz:z ;^]>W3κ;gku;|?B4nvoDI+"c(P+YSG$m+nm״/}>إ7l϶6 ML@ ᧲䑺x:d[onj=Bg mU8}!^~a[::xlj>l7("zjkr7Z[mzS5PA UƓvAW#UA7rlWI՗oOv=,z~Vn3;0X-ӅdҒ›;Bӄk<(ڒQ+ ]cm;l:h*ʀm6@]RUG-CAjbqm${,5^ ol)V+~E,,AAظ>4CPK&͐ȋy!V=͐t!`!ЄV jJYVw,{)W)S`)2k VMպO@&[A_noy嶔~#,9WV>]&eنL- ԶySRG֮21VRLk{^0>/?=<_= -΄*@#r|⿷Dv"G!u7`` &@ $HY_:@TiRvAÎNbi TN,tS̫>;[HKi/KU%'^/fZe"X D*'O::ţLV\moZJվlDn%N5ȍo.g,n.yk1%tl2bkW*@ـ'O`Lcj[ lكBzɳ5/^ڏ  6~dr$?}/M*E@Eƶ<):O]޵FY5 ȴG}|9bܴ2fw?0^FОUc=Dd/ q뇾-?(c{1v A3UiUS+dƆhC }ԡA:c?I1X;=?[IKm]v/_r_ޔu;Zq.>>jLU\,_ͻv+囿KԊFI|33] h&_Z:[![4Ik:m踸\ՋOO=țƼy߮ yEgh,7jv3]Jȳ*mUbхo)SAkL I߷;_g-umE۷HH2;z'Z?kkm5$q4>ّmYη;79TeXt̨l&sꟿ}6ۢ!K6ݖ˿.7Ɗ"Za[g/8^J˻S%)>ϯG$Yywyd ks@xE-_]z|2?s>tWO?GTϯjqp~1D3nw]LnBUR_~Uΰ+bE#b3ONhLXv+*/aqЭr**_*LtO]. +RȨ)4v[OɐQ[BW1ּT(㹤Z |W1F_z@Oz`]M,yA~q<~.U1xp{ZO I* %t`."כT:}BVIL W[&+LJLW6h;Dɸ"HX0 H{\*bmZ2 WU0 OW"Pb^S*bǕD̸z3Δq&qKc[W亱~jHE B\AվM%لJW"S2A+VfE+Qaf\MWhcjc;s31w]X/g?m섮vAkx\t¸sU?ӟfG5q߁BB2s$%HzUOe9oo5yޜq^l1o^\^jﲭ?\5=o*õy c=Uk:2[>0̟fˏW#N<ko otV~g+t>tV+lBʮֻѻ˺-!+noz,@GۚƃLk gUSL m0:/l{\L~H;DB˽w_n9&ȉ\cRqD;r]v&ȱ `uJN,tY.!+FXATם+ǾMɺ>$+29%EJT7E\ .Q{ g=}DD) AtV?\\B*Aǎ+ViCȸ CADz?XUɬ]ZRtq5A\ =z.|>hRAMMUPQXz x;9媹9I۫S c!c!PL:7"TD-F'* |/>ȵ.\Z>OTd\\ޟ)MGpPn+őtSFU/:|վMV%+L#'Tp%j:Oelyg\W&.!\urTp%jǺWJ ʆ(= *\\ D-q%*θ @oLB>\RS1f\MWpłNW"תTp%j]W3+I?C)G,R;Dڧ+zqłOMWD%嵫INgQs$]ym@ݽA;(!J`r"Y^NiQI~;Δ&7*{ 6v%~ra HYܘbe\h0.!\`T DqJԎOelfW9cʛ~!L%Jz DYvL h"Z6M'KF4r#ǘiHu"$㣈\T|V{tpET꼤2EYg `YRSq?3&+g +g>Cu%j]VUqlB!$+J+Q yjBJ HW"Tpj7Hq%*uȸ ;~Ƞb^ tDm@2+o3N+'X?Avp%r+Q}3}wHLsEZQzozLGz >':̙Qt ~;;6}8S`K)񼎫~rǾ>c`ۆ~* WBվM JW'Ⱥ7DTp%jf\MWdɟ =nB߽b^V)o߽T9.{ v.VJecmY%(̳g[֞2Cqł  +;w:;D% !z J{ DnTpjP.[WS` G +J X;D%d\MW荧pł)"qE-FS$*}v+NW"PbvTe\MWИotBjݹ&hv]aktwzU aZ{H"7$#jQGoUJ3'6)툲`Y&nQ29[\;4*CL֕Wf\-}+㴋"ks9&=LM@A}{,n}|RrbQZk"IGRQX1(fe>Ur !\`Jdp%jbǕK*ĕS ?D0R2Ղ޺2&+`ߧ `=nNr!Jƿa%*e\MWh+lOq`(\\kSu;38E\y)!\!\ܐL87r8v\J)⊬:zxbv`ru2kWE JWSĕC#>gLs_3t6(`^Rt`ɬىZұcZTnXөcZg:iF`UKz_誣}ꪣtҕ1"%~O~{*J+"k+kH}RWݟsBWmxtRWHWX%ULnouj_誣UG|təWY^f2Kw]Xo=ƛݝΗ%wmfmtXn|vٛyfC}/0]@m6aYß ]J嗯nutƶA?ɰÛ~Lz=|~^۰}=]|^^ Xvbdo0&Z c쿞l NJ1ߏ]=^~}[;)ǒuwmn6jgo)q6wū>Ƿ񑪺čx T{qaTT kȮ2/3p@ϋ|sqS,g, qr3^\cK}wwt(_ۨ=NFEVƷ*%YT6B5OU&meY2j[޻~ oޯ#4.lڋE-o10PˋY?_hU2 d%P-9&'RF⼶l22 Ib7VXgߗJ1UB"C q1rѪilX$5'?I?6(A)>> iK!W(*PC8|[Ul"IZ2 FRR(OvJX+"TyᓗւUҩd,FDU* פnEƖb1d]Rv-4=5 j,Z3&u),NPR: JUrGC5Hf,cSJD34f3̍NYVIʔE8_DD0S Y\9edȲ8I !Sc Ӝv ҄&M :J԰T`̐)W,4XZ\$g1fk+kὐ(O Ј "3I4^X狘CFc-[@!!X2hOBU |\&9 A҉YUb9c$xNTejЬɵ STBf'@xh֑?~J>Aٍ %b'X5E_G8: Bяk:bAZq m@ki fJ¦ VCIqY7M!!S`6ՖLpRTD@ (X䓴ROZY{E/3携VbP[bӮ,27F L.9(6Q0J!B QD \.%Ɇ0ѡ^& E2Uc`!L9%Xr E@ #ح(Pd8gV7 /5h;FQMJ"Z TbdKJ̃d0)OZF%"[M`sD$l ՐYuqZ|jkV}`tN# (5ơ(MհϏ@hgo\i%E/Fآy , R M a[Mnh+U!R%b5e #< l3ڄWFZ-I)GKtPGBFtT-'L+|yX*C@z=5(!ȮD@$ 6CA5ad\Ѧ!,xP C9%X !@YP&9ih$mfLV"˄:}n*oMI9$(X,xN`&BC\9N џJu]k+ 9mkL:4i_ J-7Ѧb:%t4oCBѤ3K-d~CB['oА:]لReFTcܓ. R˚[BȼT} թJ_JAQh8hl,@@Hv'PUP:o3V7_&M23u&<mf1!.E1kP44_#(*SN.NR'_0{0,bqvyqZ׋y9x~+şWN4mKF#-}9/m`-d^#x:p8G&@O[0^J.G[K*d`1uLA-EMp`CEE ڃZRH&5*2NOjEJv{ ӗTd2jnu›!q*pt/]T.'侨ϳW5 VwFTR<&K΄fBNAU!p]?kNy\E4Χ,}'XR OBFP"GA]JS0cA^L @oB%\}z З 7,EpjI7@R@"4ePEx(%'&AbDGKI(D׎a:Y@1^ |6ٕ;hLMD0Xj!hk zӣ$O@Gd⢟cn,BLJ3%Y!?A.jDѻUբ=`y(? >&@JY"JȡhjͥV  3]D5VԀ,.e3hm@T 5i"*x/tNQP&mw+$a:H` }AieTL CbzihmGyᦣ%@ɼ >pvu<;ε3F#A)nn`3 58)]QN> ENeQiYkZhΣDjAaKDo1&P3_ @rRZTHwCRD6HmJ F@bs&IU.j:RA )aIP HOOW,A7mͰll+O3 XQWT!& }rUpՎ6r%ғeDy ažbW(mQFRSdGf$dj"1xs\:%X&W (]IBX =4v*hce?fJA*Ei J>{jfR &C@0 >Y tStr|fӞ4Pcf1 *";i*C :mP fxpa+EXI&E4zʨu!1\?MД nF=>dN ZOQ!,m(TLqJ*kE!8Xg7vҘ% K}`eSFuLf5drtFH8DqB. #P5.7wk_]шwTu`B{ R G^F^ܴ\7n 1JsA0Ϣlp6޸=~yas<ͩ5^ĹJWc_O~.fߝm$2 hb右͐+><Գn }7z\xo{z^rqh>~[rbuswI{Ӗ}<وp|ral翿ٻ8g=&6x-\Jw1WZ[k? ;mw'К0^(@/Hv@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b';~gN HW:.5OЪ~f'g 'ٿi(o|p'8;7.^ 2@]/]Ռ;M̜rIsP96Ro-]WonHA>HLϖzE>_y a$ >HA}d$ >HA}d$ >HA}d$ >HA}d$ >HA}|>H iO>H6{' U_(akAvif'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v: ON &?N^8ֽoDJ @*ײ@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@6eq2,Y L/ |5-K$ҫd뭒UծmJe z?;^?^{~P\2>V/`}a2pz&&J۴q kўK״#TR%gƥqESU,yk*լ-thť;2Jջ+޲Y6Ӵ?I콉7R(+!4 7fѳ7X9E wOn?X8-tΰc^N^ 6q9"q>Hҍ4~jwZ,/Â=N9h_1 >)s>ס$|כ|?>os_cԼ~^Z2ؐǦ:e]˞Ӳtes*C.t0\SE *L՜hלFe#EV'0J@fVDw>0EӷGA(.(TMQ2J;(9ب=3h͔JFh*RWK%Lf3`[CW$!52J]W&o뢫 f}=__(HWy+cA=`+d[*ՍWWe%ĪCWJ+٪ UQW-UF)IGW4(Ss < ^17"@P/g5O`#N;{.ƣրos-(ԍ D˧Wf4eb.pekX:|,(t+ ᄱU,Lk*Õ-tǘg1v!8+ GSvǡeWP0#wtujSee㲻H .wVhaxMUŔ]lw'2[9%u)<1tʉih ޚ6Õ-mFdG2Fw8⨣i]Ĵ2\BWMi{Q:z"- eD2edSWܴ0@{*},h54e5gRGW$Li sҞ ђǡWW%t=~=tjJ[CW.oUF{*4wHWjDW++6xZxg0[,EcCR{QaL6Me=SviD4~A4dvXlzA2,fǡ 8MGttuj/@EtKpn ]/UFٴY:z\T6D8p+JY/Ż(m_havnN} d~4c`C>{S/8F{Ɨ\ ~O!{~6? Gj籨3ڊ6gvqW&nwyWz<[O>V9ϣCت9>~e!e$z'뫷媭E ^>  ooLOƓ'95ϻď?WѤn}XjvX-ZXb!vtPo8|v0"9Ss(3O(vPM, eF\ʢO N'w;bi=%\t09kuiEsWw\v/aJk>؋5,k:Ji(&|?w Pu#^Қ< f(W!&˧TvCA#:)T3G[OTxR:,{QBm` j|.DDߵ&ʖ*ZȏYDÇ+Q̵WvCLP|5qwe0Ď1 SJYa$P)9Dq魠&Y4JJQ#e1Q= 9,h)D<vŸ@wt"85wc^.(Kp2o+䘾ydWrKw?g1+ L iu)@ ZQjBHΑdSH2Bd 6X[J؄$U73g?3֧Uj,T.5s\D6}v3<+.vu⧃?yПLqF @# Mj 6Kkp|"bRXP`c5g=12mr'!(fJ67`d1t͌]FPq1ڵYǩmjfmӱv`wZL9Yh\!R TRD)M7y@ zQR@/kB$^ȴ(Bxtf> ~>,8CF>NdDFeyy#v $l^/DpoPqO^hrZ,R S+#M)@8 $( X<*GVQ9~͌X|⤘w&uf%"iNj/x-A(X'NJAl MR:^/6kṠ4>-c75:Ƈr;؊e9b`^DwW"_"/bObEpK, z< ~R#gG mֵ>Åho^Rn/; M/]*_Ttnˡ›ױ .ib_XhH\(օg_(<4;<Ќ؅³ORxwD*<%dJ^'JMY.V^JI@uH]Ou42QǸָzDK VH ܠy=5><.\J}}nlzmqv!޳ٞ&;z.`f; "s^lj\;ٲD,llRi.J]Y_es=\HC\/x1q]6 }C}jG&b΁mOU4-Cs[ڬ!V֭CY=ֻ]RA0LJ*& '/ s%F(WD5sَ!)O;HJ78 e`ޤYjHvK;7 ;̽h-3e($Օ$bjQn7~Z|Gs1]h"uqf I SS4"tR)UIPv\):kտ~/d\+,y"RVHN 0`%ѨhuBu*H\@D 96hf(%3@p&e66 % zV}mpO!'&-é{}>o xXl^68<,EƒgAw[6鼺'9'ӟ6*\L: m`^J4MD -r@)4<1n#Tt;WoBߝyme-eȠ&whb< opB)?J; 'l_R2QO6H4/d/P(%VeuihhFh"GYJ TA$:''0 Bc\ʈ`_o*3!(\Xύ$!JʘT&H@xFIS v7+ZݮA{VWfs9SixX V7p\ZN\qqܻM 2ӣi-igߦɢ42Kӏh<[i[[Cۄ@]?p8Nߚ몏ei6FH#\,`JBޟP}ӟLŕ<˚bGaNP_yX)UEh*LG݇9uJ1}nK##IFL}X<$K*'i-*k$]GDAgmp2rl0`R<"J.d%D& <2ӊsK+g׵9[E:qD2̩5E}É flT6JLEPjJü 39*9<9CKng[mwn$hѰNS ggm@XD>h-\T 2ʙcI[ c)C;4jJD;5&u9*ɐ)der9{2O0*]>c9gαD#JNPcb)ĉ!JL1Wt) /z?cT k^/dW"ё4K&- jW3Y8z Qo'v)ߩ{BQ;B!E9*NIڃg3 [cY;T &+NIw~;Dl2 XVd`:[22Vr8&#F^ںm%*]L`MW,r𙎅olʈ^9*}a%9%uDR&JzN<`l]B39T:la?k=[WȢ!U3d];kD[$qZ#ٹO/|QH:r%J ++nJN($E!)QH24fxt!as*eRqlgJXNHpI I ,ctd ft-JQU_JL@3PFڥ(x83rǩԣd4 N s,H,eb$~dicRZ2YĝL3)7T= `J:jh+$6gң7 Ȗ)Y囁.C6:tBNLHi"1p.|M/<['i.0P6;Hu0uu9oIruuw>ith>4DNz!( iel.HEdD ,"G~!k;:Cv3w;P1#egRT^)r;JtR*W*|WyyHWӃGmuK DY 4d ( B{@Gdm,hPXwH;'̩%R;OV _IA1D$ -΃+9#mI4eȭ\ljk7Hթh}+xw߮%6ѷn=:z't$x|t4 ԕ#X\LQ5 QکR;/*K6SDDVA"F!HLUh1EmгH5h,#Mط)AZ«8sG'0VMV((+wBA7 ?T B0SCUP+J8jJjR!?F'rc7^G)UR9(e6o:'*q+V^{%]t荷C7jS5B~eݒd-n8|72d?ZK{jʕ]VvIN_$FHA:afJ9Ib"Ιke-j9uσ .u{Uo/d;b4(Y)tU4;R;]8 ~@)}onܒOThvcae=2eI\wn8`˶-GyP6nӼ}Ozkӽ~[{V`A zz'!,`gc;YCHRo}٩j:_^wkt`bV+&T:Jc#h;O3G(hPwtUwם])8)e021ck}!6?AͽW 6@Z#}AF( dZJYr19›0Nd8Dfv-W]tFΞ3B5M~vJGhJ;t6]xbXyev=!6c{[BYw$ K ȑUJ@dYXN97I4\gޡIHR18I f丵BhR:*uy^w@g<ŭ}Xl)C ,ɲ,DunYNnƙHdj?w`G2LT&Dv$ $233OjGSn?s'@ ޜ38̣RykBFTI:EX;!(Vp!|cͦtJ+ nL[{ڂ^|~$Svf{9Y}vv f@Td#9~G}/Y\f7F8F(|b-tz`c0-Pg BQ;Ŝ!.ddbdcjVeiBʦ"yi-&h'HڥLrKV߭|Ered:}Lȟ';$tY :.,Ln9|+rO/,fҨ7V~[aKȃ%%ZKr8Zqd4nR7Z9a֍ڦxvP{Mx] "{x#FZ췎ezb>jؗ,-`[\bEo]Ե2Lm3!|oY|mSu蔶mJ{M4N#x6n f x'x"iU"iu{.\WՃ BOW{>>p-}4sϓ?\"UR(=x9zngK+Řx8L TZ &hӳӰq̜o+x2xdxX[<{]ghLOx3 ^ ŤoOhe/oLbֽa"hC=6'9g4،,zch֒ w!b4f'$zq&RI2tUt@O+RH/R?9~pk8YӒvִS˚>){ Z"v8ÄI!db(Inqidn7w n_[Hr+fGmqHwj|X|JfR&)0|&x,{9 ~[=0~ݱۋի>ߗ?Ofv)dβMBN:xu= H r愊Π'&\9 dvڢy9Eoo(ީin9{z/jf$7ylѴnXŤ׍enVF =5A̟+sI9%BWcS[bM*EuzM di0YJw+ml@6 dRF23!JY,IEP28*x]횄B/sT3qvt1顀ռg"ɐq'q[ Eô_~EY4_n-)mIQu==OgӢ?4Wz}Z#QfXH-C)d3\ę%[{׳%%VL}꥿"% U FGѧD>#5$'kGE[+U5|?\)iyNn8U?\o2M (HC^QHE-SP@)$$D%ił=p{nhlj=gds-SUSk&4\|})bZ ]5o38d\ 7i׫.i>EQtςqPN6^m>O~5t[Y+<4(M73a:}'vdk E{efSiR3oh2̐sNZ(]ѠK9|({vNحhN?V9WHCerr^Q_mZ=|VSw|tVVb3'irZ)IDctv<3̍dң\ۉ۷`x2BDнl?V|7/X:ohzRDK=Fk)Vŋa\53$C b _9r=rʩE)Kk5 2瓌^3AH \0ZhκKR(cbY h +nF)Igq+JO&)_;elK pW7g+d4#_ i%/}Z3د$.VB)0pB?)+~4J7P^_O|Kiب$uCkA>VZx[KE6_!nI/ǣJ`HSBȩůIZ=3B:ZٓXKBLmxRoT1 )wGqoE&ٝ"% h6[RnGÆWޫyt1n٤uð6Ù[Gxsӫ`Ǧlik,佢Md 5SzD'G;ߘW5єt}t>}ԝ.yp0ߛAzyoں{u-~9[-=y>L{[ܽ?~λ;zL4r;5ba1Л7\jTЪ MhՕ8zH\'Gei}fܺ3vu؋^nn,_uf+° 2xu&r9Ȃ PB'ۙ. {I`Ҩʃ yƩ=K !D:sm$ztjge/@ӢکB>+!LE0v4yYghgw ;z{C!B ` @Fd9rR))՛@&[Dnx/!Ho)c)M Dɭ͙Q+91gmfG7@;b6]`B[hif"2I]T"8a k!# H gHT Z@wP9|]jencFs6 \HkFm8OpVXN_DнM0fݳen؀56p T=Y1yrB "Mgh^R\M@BF+艪j/[́KřHgeNHaЁ R2Y(: KGTd*$%S{e@L$ST1;$BQyYݝfW9'3K^Mt B{TE&|LK['ߺ8Б&_o'Jg~hץP\I*#sGGgyяtO~u(k/ǓqL 'K"~ZӟOZ`O+Vq/) [)Zi[:]d[#MzwK%N| N[lunBw eҒWϑmg#nAG(k@KAj:^~_m;?]\G/X ٻFrW~M$/ d$ bX-y%=`{J.)[.:wsi/=dL6լe+m2ܦ8"ir-ZjJ.ϖ~?፹EY#+DUQNt7yѰ\*!+m )>mYpFE Y6wU[:ƳH6*n.&Ã4,_7kBdۺvJ'I2/wH˷$>{f{72^K,5-qk+d݅ZJN5@KL#ơ FAZm|ug%i>wֽ.Yyf>͒SN%IӐ{(dK/,3\/~E '~^CiȰi^NnT^LkxO+ө$*Ǜend`& ‡ FpLcJu%wyKVYP]JVڤQS̢SFkd:BY(c"Ι+!k֢2ugո[CCFw聯b{_odes[{\P؝ Wl5.,[&o75=+&6gV\3rsJ[x#+\-k[ua Q=Ԩ~v4wv}E024ho$3)^CvYmm܁Ǫ'#Jd*6?hק tqc@А6GS> rDP{̹`&PfdBZ}h2e}L ΄Z8)J12:'^?.4/_I Ur\ίA|;v㘖[Q!6jqֵYGv~jiTdIP3l8KIT\s*GT7R\eY,=ƖR^ dUplA`N#"qT99q g?ȳހ2y!͵W_Ra+ޮ,˗,Y.'^OVꝟ[lO.H&`FȎDѠ2+3>: ܪ|v9Џ>WXϭaѿdLY%\柟&mYX+%d#9>p^!E}bD2\pD(|d`xW8p kUg0!"*' ":9阹B^rsndb&ey$ce E B+'H<7KDvP;pvHea{*-I'g;>8Q)ھ~DtaR5\7%avOBe`4^h!CqԆHFςq-_SqLg8@Z!p*1 &QQ:b`g^G+DJp+,(]Iy ]wa\+MSV䅑kFƉ(#: ޻T$''$"; [/Uֻb?jB QewAVK5@(p.B3x0n1*2 '&SH,:iʎ%@GD]WNwv.3Kyb:5#I҅4 vgϣ lu"Ktzr>4dWZV.{rRnVn>bzy@~zCPCi O+8=_]aV"OyޜNWˋӓ0'䚟hWf~霰EwΔ)\nt&q9qhBiҦnhIDD4Kn{cнroo8}A]"h5h=G P} 7Xb+"u/mG=)+)k\u=C0Oe:[ץviub.7 w|ku5(;(C hxlZOZx.i#IE5o>~:o۵M37?_hm+7lH|bt_9?Rߦ/MUaD ܭ|wj6ont)ÉʀR7|LM-Q=bމ .I=v_7߮] <2ޘQyW*̗_KO P1'd@4E%P`r{Xj^ (;p:WS7-B+ÕtBVg)\4}!,$)A68W_I7Y$H`F& %vF*t&rzN+ Ra"H9@<6zTfҿ6jJ5nj:ZE.&<ܓ7P.Ue|t8VzҁRI4-_5n)ƶ1)Tts׫/Vq^X֮5ֈ@xHb;0bZHiC?sӅn!n) kbzi%]>ؽ?2tiS҉J,=i|?hti:H ҋvd)^z֮ws;p| %N1 Ϣkih.'[߶[Ww/nӵ՗VzקQn]?-Q:,?+CI,fyaԝX]\(써Mޱ&L㈤q雡^Φׄ˳QïOI}W0v:}wMEi+Bs7C S|b0gFFͦiIQQ g#D>1h-nTSnmʫ1]>~4.vJmY'FKXr#Ɔ'OưAmt#*OCIwZJDcsxK}8e\c)\KT5*d2= P\Ġ2bUKIMr)hM#w; tmЕi gG}L1roGb!__GVJCs4qRDqU}̑DqMiPҧX|3L.ZO2`[P=fhq/kq ]Vb $̖ QRBI;&QL׍ihr{"@j\z]t&E58Εt@62VvdUaa5 2 i&G05{h}]w 7O7AN9zu 0HY:hCU$RtZ9iKU!cE{&hZF#l4l;6 YBt)`2bWvI\11IǾm+P{`%(aQ2dL+Cdy̜z 7j&ҒauYiG%Y $+:@251E4G(18W g;F2 b-& -l"C ro#k 2ԣg~()[-rR7#6NJYH ʤҀ3\წՆ%Ԟpq̯zL{դd_\q8&1A!2$ΔWL |?"rZ+r$^Dx&7Zk4Q x\<<&!TCa[w5rn8I;77}'Iԍȫ{?K -sD' jed; F4'x >Z$kx G{};^c|`I%W:瀶$d(;PsgBLs"TK{Tj ȱ;r<폃m|ޅE+_zCnZOna_ou ;}qfJ}-(s6Wa 6-Yj,&#@?~INܝAϙv]n5yЎ)qRFypB0bU|20\ͧ Ӭ;t]f%0?. B W/3#%w\zqeV;,pJ#)'W\TWZWM \w7Y\^-e$oi~%LGQ2gA z6rUY/svNVxa ]œ +Ru>I-=Vm{WQϭ}/@芆AREUD_h2N{'4ڊdFf?@e>6+Gw8ZJ OOWJO#\CG2`U3ͩXk5 }Fz?pWj5s \5k=4&0ku ^M/c7]ga6on)hUNM^lJW6Lj6PdYy*&#gԞ r:WRWP"jc-d9*b3WЮBuJ: G J+[Mߒ}|C=ŵ^ғ3ϫl~:ڟXͳJ*ApG_Xxfs'7sdfZ|7+7jǴ {rr}c5EAD]VS̸=;Nx X7<'7 'דS -^e39b31q**eɖ-+XTHY9QVկ('G蒲 gc\V)6ƛ|o[7svh':|MW1rY-웡1yV52jKև_U 3Ȧ^]P)So?h'y"Xjb'E6NES9@\T1hk6mB Zgy"{%DgٳSnX0@`UA AeFgJ)gA`ƱEv^ u)$T*EbU`sfΎq)xvm% GoK)sZBW*٨ۙm5U.ZͶ+%r2c7?z>W`1@yOx֖Y\+.!YCe ZT?Wc4I{xm{ήDa0Tr rTIuG #Eźwh_10p\Ǽ,\w_>OYc{~ԊbJy2QL [wt^ҵ9"D. N9`ifDĶ/"?n:y_o.o𹞃oOh-7IlL$;A\F l%9u)5l-_xܒnH=){}=N~5nyJ"7iw˥[DUʭ ڸyދMŘ fӫ2VxJR.>xU;M"GSހe̚ǒ[*I$qR\%V*PC ,<['9'P$Y.,t$X-IIr $o{8=l=G!F)kUs)D uC CFS{3y[2C8Ml U$3qeeIKJ20#)bUt4b^)I`9"^R1*[F$yJA5Ag]3wOٱ;dD Y.Y}Z/m#h̷`$Su2<)B67P ]-YFbtME-ޢj;;rʑVy:c>LJ^u6%&bmՄV`$҉}TTrѦ.fOؒREZWr&X jqPk19MXD X/dF]",eLIdL1X,cDqq53k &jUZM!"9NZqQ?C1R*iпrI -5fՏ(CЛ M!)StOLόsG8@COʺvYp[S0T ].gw<AwEC )"F|H0ʒהڭSC˞.7:Ig_/ξL#~:פֿ?ہy~-ݸAǽ~"Jr@$R. X9C.ӵjj3+25oeV86qhs>nBi*/Q>&#_tS-6X~1!%Exvvr'}8o{ߞ''ӯ-ˏGxw݉^Rmg9>s/#Ǥ7(iDo~l[lǻ{{YZc.3uë۶Y6?R#Q=O^jD vp2OaiJflDz2okH6>x=Iw}=CJ(yv%vڼTl!WJҘfêAlv18U.XkE9obƈc){ҐLB# !D"f|~zp=Ye//;tkk]휝J/T|FBHIxSs&gT6YiefkKx5H1+,_íc0@k,:VJj {cP@'HPդ5͜#Ql:]mkN6vOun)bODq&H[m_>y|^G*FUe:t"=Cwh~~ p+확zUb:;?!?ٻ6$*tyD^=z{v;`<%(fQvYER,eSF*fFVEF|覺S_zW4IK6.GPK2vsq FM$W~Nyn:?'Q(G\~9 ?\䤬^jzy#{"o.FKz7Ivq#hV킰P69vdeInIŒ.0 !j.(|.qjv[ I}uwd v^1ij? Չ. :fLm./Y2z) _L"ޗXS5f M<0f6Pnx;_"pu̷nH  NCnf61Bk~ȡ;rؐcϞcP8QJ#M5)syfϋL  w'tN8Au֤/)K(a1kdU0.祈9H!FHBesŐ%aG 5Œfl6jObF$W$iDR9{zE^@.&NoRoSAӓ ?BA\xs-I ʐf7<x0W GZ BF]`D >Z@@UBK٩7bel`+U^e./WP`YE+um8"VI:ˇّeWG5.+h8HKU60)yHhq͍N6p/3P$Y4<_W˵'L5 Wb_žAY6!$HPO'A*`ؘՉEV'6J@fVD?}%}m:^0]8l ,5@H\kJ=iZ 1AxA F8Bdiwa6+\S5,|]%56z!Hp Eֿ ($(H9w5X#kcCM3uQ_UUFZ|7 ,:];az`@ײ=!BZu9-Z"znwʤLeE~iZ^HwWD^W~]5M[%xZ#MD sf\ lҲ tj:pcm~OsSС!TZ*YY5YO7Z0q<|r!Zz(Ӗ(R?65J;mA03Gy.m^& q*FNSc.w]9C"e`T<j[R9& #g"R࿏q~} w^HA<F KSM j^ETyʖъ@aHG%d< gAO>MhNѾF-Fը'$qI(7;b|op:trמ<.JfY˯osb7M>:ּv6p4ϯRy{~ST&v[IlGfW*ԜFi X@ hT!U !9GprPD*$X1$R*&R%c1rKrX,,B]X,j;> W.QYDڑ茍&m^饵 H>htd1X),(>fcFs,6T^3A%6`d1ta]ۍa<];ڦ6'}B9^ ,de)厇*)JC&֛hmKJB*BFdHhϣBY"ED2 `]H"FuTtblQ?/HRDdDܷvI"~+pLbKN{xF HS+LQh4m(S<(`)eT=Ri(nϜ9%ˈڡ\leYK\'x8zKdJ3xD8AqXNx& Є;_I) $!E}!/,~"lK-֮l#z_1+C)>˭*Auqq &ڛ,L4΂ WmWw8by4h;q8zbw܋y"? J2%'N;4H\G1IP^)n6|o@7hdqq,DK VruKJR_Axֳٖ /o.r+9 Ҧύ0F@cFz3z]0]29«S :K=zNeЪ hKIۂ4KlwoQ)eLoy&m&|ƛi+}p<}sħ3>ߟ'Q+yӨ\EK$)k{9TL>k 9 <[d7@͝?dVvUΑ{ȋ<ݧkeLݯ_Z~:ݪniZO0QU7Lo'æ}?V];YaBWὭR_Ŵp#*nt;Y:>\T3OooWoԪy=C NY89$ =PMV$dQ6,PTUB/}͸SGq YeSk/{V BJkBUTD4( C)k5Bùl!"B31RH0N:\D"F! wKPT9;&.re:K/SrkT)jWkb痷b6OxsJhkYޢY S 8-RI<8+%uD8TT {KLd"mY+q%^?~p䴵joB2 )R'vSURQHh(U#ō%y(q_[ʆymnkbUf݅:2#[dI(HɜwxÅ!Q_?RM;m$ƢT4͵fZH#1 OYK0kalT洑nEZm' <:|q9CQ&#6%E.BL>v;G:瞅V BѪ mzΫ5u5lI@j 0シjŰa'wsUgJOz(T3U3F~Wa]Y콼ٌK&0g> [LXu ~lɾ8FZXLqQvWKgu)fl}}ZxsA>KH{O5!u#.*\A5>=ogG0NCs<08,!<"Q^>#  ; %9ܟ٦7npwLQݜŁgߵ$h1H8j=!٘Pqrs?cYJe */㧑 l {ǶU{|;mcML<4ӷ4cY_\88c!S)3_.xg{;gQ%g~;F'-B2' ? /?m'wnbweWge<_OK8gc5;kغ9.3mу7TPD 7FUG|TA"gn% βp{n hlvv66i JC Z4.73 uúQ'G\'IZ{GLZox;kmӃE稧G SӋ>"%uXy)GZDH]5 8g0?E#)~)0^aLȂ'|@:EH" 2P E€D Q (L q@D q hfhV Τ&rT@DY@gaR1r$g&}3?>$B=/yU>8luB8w.5ݟ|ܱ M<_6V!HbH-ID"WK6rNiټ35ݩV:T$\e$ %lm>bWi'7Ys#\"B\"Xlv#[ |! ןcM<#$!vGщZj4Q/Axyd)=x'-(rmbLˢr#$>j#?rC,H&"88M&`ũM }SB-TJ(V3IgTyEd:h5H9CӰ? 1uCOG$R++ɞWn[tpzPdtٴַͤL\Ae\g{㋉ =CKj%k/l!pQ#'޹Frkg4y[MFNr 򐜗F@<"i nsiKQ\Rb7HbRJ~~Ћq LY]wLQM + >X6zKgXUU =>^ℷvA͢V?LB䫐b!]w˭|BFO/.rڼyi~=}߬,ڵEg5ѢVm-KF_u]$|eyí7(ڍz ly$yrۻ3 "X%^_^~n8'&^ǯ8r#צ]p'[ h3ox{||}u2Ox},Iys43vuͬ'j_O(#}3=ޟ(lĺf]BPMGw,kѕкǮқYWG+v*l1^ٜ$ {éE3eZČ(w064 `yk^V4-׮itvj:斢JkgNpC3sb@]WBy7gŊi)|'kxN8ZhN8Xٜ8BWqsXݐL*6+؊u%ͺ:N]fu*W@Y5 n uAϽ1&jcCp`ߌ76OjPW6zcZkGWe TD cԕW7+kcѕjEW5J(5ͺ:B]%6+|ǃqSx0v] eYWG+V5+vԌ7] mpJ(k }^5MFYU O=5$V(; )4ci<Z>8[4ԉ#'_a0ߖk}0v"]ΤYWϩz`lklR|(] kh;^IWl}L۟e\Pd-]+QT3 }3;̌dr\1:7eie]Cl(GZ)K(u+9FU{J1(F!]8Xߌq+گnl&9mh{[\PԐx5.GNv8Z~,ueslIW?);qz]!Efy4+eѕiEWBv] +hGW[-=p|G[ҕnEWBmީÙu* >DӒgE] ֭Jh-ծ+tfꊝP݄N:5DYerv=AEji̎)hZpc3-MHJ;'Ǩ-. `Ӑm+`!]=Z` GW80F5-(k[XЕuܪ'VAt%A5+eՊvG (G ^IW&rݻ)U6śc='֬ʵybM;jL}۞iNmvmFU JqCW0o5 [JZE%Ksr9 t%7+z 8Z~X(Yuu<)e tZՏ# =(tInoR&UV*,|&ZgCBVTW{{w7?5 6yFo(ڟ6 j}Oe*w/}Qu YU꛿|'.ݦ|?z"VB:z]5M:9bU!F[%5UnzOfai䃿T1қS|f|~?wwmkSn6^KHkjywB*T~Ͳke}XgÁjϸ6_ė){gT= w.~x"A闥j% y9I[KhCOMIsިS6:ω|GX|9'n?n4g ڇk}^ݠs?t+?D1PQ˳w&DR_y;rLd4()g,涖ުg}>)[ިs \BJ1*eBʹ=XA~)x& H}b?iP])6*h*uXB@c ۻӠ2LC;imdR?$mhgdrPI%P( 9E>dwҲر x!ծ4 u%k @&g2")]O6J 1D2h{wQ֞КcNCɈ,v8t9%C ^'Z41h՗EРMZ)ʅ 7䱦2V&]:NrPUߧO ϜʙT!$<]PZq?H`VChqxW}5-h)t[{u"+:ҝ3(y uh999wyJ >͝$wUGAy^ԥ~b&dP6FW+J4HSO]FS  g:d$F®hA<37]wp2#MABXv@$tHX@BCYh]i,ZiQF? :#oY!NXܳ? ˄{Gq#B6t]3gj7é<"`z&_w^ {T\BD.mN}NXd5h Z V):?B; iƭ+x1h(A.%(Sv$={Qɧ[6wmm[Wc&m"  ry`dԭ#2)_fImQ֥mD6WWڵVU*-5r"Z$VCMer1A!aUhޣwX>sA། rI+tR̈KQEDtNcB*&6/ Q;4')!"_ 6`V9݀m tjl kvbJ;k-evգ{]ܦY+IJ%‘ፍ`v(J)h+>$V*X(TH9 0X!̠`!39yjN$rDV z^U 6!+zxa4 OŗUe 2inuG[@q{ 7|tX:?(j`h%hwV4<¶AMR tY+A";~ Ɵݟono)d}E>j`Ye>bM=AKR=R{ЋAuD!:*Kt]H_>УLBP}ֽ$V@e*D-`JA{ AnK)AKcEXA%HEWPPl5kh6ܷtBPG%!k,:& E@iv%&P?=(D"hwEdUU%o  4,,{cF8&gو^ [ɉ6łB;gv XI#,IC5$RYIDeo!mJMުh{9BEx j An%BG Hi/i:hI&hbV-ummˇzԃv#[de1;dY:VӴ|ֶ9ב2IwU#["AloNN"ip$ti#ll%f6{O(w%[tum(T͵K8O]V<8J&h31irzPõa#ҷf>Xe1h74JxK-Cۚb͡ xy܈hBD{=hD/WRhAvPXfR 4Č^")S#1l)b*'Xѯ(t"Nd@'7 A!gоiAK- +U C 2P(E@cD&7cHbz0^;:P!"[pOҕȪTz19Mk.hYLf -4֠XAlR>ɗ U9e2 d2R 9'kKiI;:ګ &xZ$ScQdmQ‡  m` 0Ҳ¦g@'=ʤh^HO38%7Ì j|dNjORS((qKZkCWԤQH d~]D…0XoCnИM5墑roVå!D40r(w#jI:p"(Pp$@B/\ mB]ш;S9hQf-qwkq|\/\En7ˆNWLJǮKwioM{׿{L_UHJI# j_bNkqɏ5X/K9lit~N+߶|6_Y覵͖:@w 684_,qbqzr")<OF WXН۽iF§wVVK7WGx}[Nޤ)qd%ԣ[) m,O oE^P@Z;<pAu v=K'P Zv@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b';Y'aDN ZK#Z-cqZ# D(a'3tѢT N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@zN vcrp0qnq)=z'P*NBL v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'3vAf 1&'A~qqA}N @q9+\+ƔIWw y"N{"{`NW']u؝'%01M(w1{:aOt 'žZ}*_M.yvRvRo9UWNGeMՈ4I׍fB{wȻw<41(]ppׇ+T*tX ytu?>hCӤ{Ȧ=ҕtЪTtTXUdf5߭bbN3՗(, $|@ť߿-ddǻ/d7Bn]f]٘?+=bؾ37q-gnx5_2̹o" n!lפ;SDͱ1*^<"|]rD-v4-v,-F)%ro {[cJWM"r4wP:+jD IW׌]Zwj2rzDrT(]z4>u 3LWNH$ة+X5V]/wO2͢^o"A<7gW]VAm1d_.v߻§\]OYbWSb)ս~崛$O L ΈMRh}ꕗZWZhQPkN^>$52Ow>1 u:1[l}S޽;iSͧKź}1D.v]j6}ѿ/?A_nIޏOiyϊ\,'\fSaguo81;Dbߦ :;`?Mp{j{7'\^ %D2^,e24꒚WU4{N ?h`p@V57j,ZYdHJGt.*#=VcSַ%*pV>J% 1)zM7/?0)缌NM,֭FB*b"f#mFBΡo6M$xGhT tlt} =="wq`/|^_͛|wlc7ȇvGa5Dw =R]DŽkC-'a4VIs 垳?ʈSGx̩VqK}p${>=iJ;^hv<\Vmt>X׷nwxH=+ݽzm!3h & U]IQ6c5MuYc!IvH} Z7ɖY[^Wuթ69igm6Y:ok[)ɇF!r~F9Tuj6;; _7уw,IOt.%ѡ_\_y1#Zrx>+s~g[h}ىmHIu+-z)S$VzS9CK,37ޣٟx%M[qN\ˢDQ}mv<ľa[QNתKZtΫ\mޗ4E+s6AT| M[e߸A>bv7e;\vJ.KW˳5wZr=r~=B {h}Ejb5&/ٻFr$W{1i> `,hл` %Kn\_?aYe%S\NL̈q崌)͹LꜤD]U:4^Gps@en@HrdPY-g?ɳ=N&؛կww[laW+Y]0K%wϨ O v$SADc"G#hdVffB#׽NH{vvy{NOBT['yRJA&b } e@ND /;b\+]6QͧMH;t/c4 [7]q~7/g+s+eJ i~̉󒾥ϑX*8. $Pdy!=0 ):l.EB6uY$ '^hR)e{H!FΖ%t}Ɵ[bz&We-A;,JlYmB*/ #qEJfmqMZ~zIsp֞xMwܘ09 Ž-L{G1s+] tʼn'^R0/+nA̞t.F,D;vCz;&Z !YvW csu>yaϋ1iQkI{u%,N3|j2ƙȦʇ{Kz{^B}B^Ux缋!&tD!y֬N\d-ˎP@Y=U"1_5-nϥr58Fӳ ?BE^xXu9n>Cu2%#4koݍ29b/x*H[ϕ6nTLq9&1V;멾|cgZ}/Y4Hzo^ړe5VԵ-;ҬXnfG>.k׾L߾|uw58yTȄJ^#S&@'3O:5˔0i%IGp?$o\\N9Y5շ|Qxh-bۈ6bm41q} W2 dB_ ..fDF{}'&XMqƚF1Y0"ǾC; Թ }>ZKN gDpu.~CJpgstķJ`%\&onG,eeHB\)"\!_m9sJR{&hHdAy)'˓9(@{%nYi"hQ 1h떿CƯ̆ i@pI[WBGT ɜSʊZm9u^-iIiy{.I-ؤ{&I-{4, $ ȹrȽH]M-_+qUQy^sNBH03IncDT-ʉyP)~ٜR9 J.z0MYdV j#cGz\VcUBc J 3ޒ9?Xi>l{EoFɗѰ}MB]`<فI*:oJRN' *-9#c`2!EMil4*fY(ʛ1h]eĮ&fb jWcQWFmߣv`wJ† E$)ƌ LYl9Os#Yo1F)bv,db\%Zš,ECđ0>N,Q2V[~ [ R}UD􎓺YE1j%rLH!: ]UPV*UeDQvYn|դX\q.I:!DtL BxБc4I1$z\<. VcPVCyPׅȽ7 H[+1p@\Jяi'G?v[ mWÖ8el"4uPҊF&hi $z+x;w\wGݱ;1DWJ&XgɄ7N&:p3J}TJ]y.p!q}<(M9 עT6Cw-.9<&w@O$ ڮ~/0gƼ@sfn7{}L`;$端W=7UKE+JIf$β ^4Q(S@e/LAnl;i=gScw{F),q2Ŝ9eJYr1`Fgp' sRa4x1KLnfB m-D.KRBI$[BYjVhz.dW*VdR>m 2\>"]X!Sgcy4"Ѧ0p&kJCHG מKO%9Zŀ,i+d>h-y 64d`XIZ樬VEdLeYrnI#HАI`, Ta8y]_MYk+! HHQd!pe1UʌE:CI2e4oM/5HB<{'N pDt&1)U̥kT:xtZ+R]E:*tdh$1ЃC<`R%UґȀc`R'/… gn{ݧ/= r4Ӗ?|oRא]ۏig8nk 4hqvx5.NIjXðd^bӥeoy(6Χ3 fi Bu+hI)#Wχp>ԅY),ӫ @z;^}^j-OF$l5bܩCݸx/ gɗZ f"a?-DuV?_QJ^F]~旫o[Cu.i,C#58[Q KQo?u3$S7E2tc3.eWHeś0hg\7c0 9rCd `I62r=au!J n>OHDI"ϷmsuaFm'-{Q_WB &,pG9w~"|t\nv]wD7y$#\yQqZEmy.&r5_{ӳ(Ҳ&W ʴ_> ׄ8|ݱmF)@(NҶ`vOwNhUII{CVQHEbfBJ!}@'.I+@-%i:m3 #!$9,Z!@L2#!ZBpiO,;t4U[j2qLp]1 aMn>0\[6.jlWo~n_yV[>e?ػ8rW+" ,}p/Zڒ}͟hi,Qk^MZݬSSh2ږ(S 0Z.qQ'E/[7#8YY"lс{@0#9xØP#1PW\-jV "0cd%wG^<%BOw"3mmcgo+m՜5yb Bc@ALcV{ U텪r3i_TZES&HX2<" 4e D?*^_.'NeV&Nώeib9~Y.զ[S69}Ӫ M s0fSZMBTkHU9yOޱYV "W<]ǒ'7 ` 2 =sK,1Β ˗}$}ki`%ɥ1Ik" R{ Iί8biĿG[X!j:z2ZhPe%7ܳs :JίG;|HIs6yv.-]8$_&dmy~fo|дE$Ng<|v_IXRm LW4'9{Ξ.~8O\GDLJr$g^=o|s_-=lq[g><OOчony1\] _Nt /]?Pht=4nm 6p?.tgI:b㳋")|Ҷ>Qշ1nDfA)'U[-K*Zoak1jzg).&gS򌎣ʩ kI=$Uy..G9-6k-=xBFh }°?;Ozܫ p xO8vo2t$*mA@ˁxCOISbDȨ&Ja 7fbSѱx:ߴRSmySԕN`1)h̀@1 KT[IjrlW<&;g %jK5̀ɞծJs 2(+ɍ(3fΖ R9eY-~y'rjc$*X)A*Ƞ˹(YK6a  W+p.8N.wr~G-M5{$];JXݓ. s6s*,w>E<وv% $fČ#Zkd܌ȨOYLQ ՘82d' F1޳UFqf@XU_nO謫)e±PdXke/!༁VB|yе: ti帖E:>9hH>/_MHq mcR"NE]9@ O.k45+H30sy$8-~<~v(S T-$ 6U#BP DŽh L+XXςxm"@fhV"32.%]TQMg?fΖ~6 T9<8VnZL&ڂ j,T( q,+KCh t?_-luRrޓ^j=^R((d,]u5V?.>\,LS>H4!aMvC6AE;d)jӧr>.x^$(>xU'ul)z]D3V(7%UT8|xakY;𬱭`Uޭ~d=ohp.rp.rp.r\O68sXޣz`gkl5k隕6ܦiPE|PE|PE|PE'shNzH=M48]ӧ 6$WD:S2ڶ]dsۣ򌎣I/dv7 UE-s :k)m@&&|i2UwDyݏBsDO'i!#aҐk!V~qG(h$8ώ6Hn-i N"dTL%0Hy3) $ifqZ) N+qRI,&&b (@$A5`D jrVGbf@ djW%9n<FB/!l<͜-bH 喹np=.zږ[)9i['^,mdRHJdZWA]E!Z [@ZIvqYu~ ?ukGIѷ3/\5f?ݺ^ bh+sG%_J8gp=TbI[O!O?:=\^it$C}g1E<وv% $Vƌ#|@sNՑnFՑnd*dr-\ĩ! ,+`l0񞭪6wF3sªVGm!;6 R8BWȓ*MͺmBK$1 %L)+Z͏g HZ,Hl`@EG62 Й">,~T k+\yլLKIWV}JH0ptnl-*~܌}D 1!q*љ2ڙMXTP@DJXV֗I _ [ꨵ'yTaV>%Hm`CI UPjZȓF&3A9ZQV|%CUɁw^r\]Ǭ &#miS1%ɼ4iU-Q E%EKVVP%c)gaoEDlr^䟯\ywEz-$Xdt$;-؆ `Hrb fRjv;ܣ{\)66Bwe&>S)tsEH$lkxʆ{P>Lgg'@IIj&'4OrEp;4 <_tsS6OSh,ψtӫ|y XvU`I:f0V'I'-O2ZS3$d><;?2۾>n {Bwx _.9﵋X+*b}?=c{_o B{Ry,wYzk3V{6s=K&Vn_F7* aۈ>wkñFkV&~?8:rD+ox˷j6B]]*>mzx?Xl ѻTdęm1)tje#FRdTwuc~<. w$"mg4W5[ZFƲwcfwOUA0(. >a)l>oŧ~q-ʻY詯E]柿ݲ!(/= 5z 6xk=vۛ}sh\mUW )8dLѢeFTVAr[ɩ E\ F"=2㬯ea63-h3 [CLa*BTڬ 0RC$)c*{P y*]i\oAw䅑ՏzGFQ.(`5V2@:Pg*J Il򐕰xԢ =ꏯ1 ~ާqdN})jٝǧ;ݽR xuioƽ1– Nxc 3U-ji^.hkz'}#vݍi]zT;SBUӖd}3J,VYiCr-}4eb:Õ|-|~R̦Gi\cLgl~RBMZsɗhzZ>8jg\WTϋ;Ӕuhm 9s?t狳>M?ȓ7;ŧӓ47v~qr%O4w eؑ{n„!PQmdi#Uoj}w;4,_ט\Ȩcg6ׁ]ب_w׷c"ߚuԵ:!_Zڐ~P ]~x:_F7-l2MMZطIIT;566#fXwZ̸s{߽G)Vr<}@C]0 OU ><0E3~n[]N 1ܹܻ~im?pw6Cܥ49:k hu}~Dc|6xW-V/}]-Zǫ M|]{L]vRUMc6 2b ^X,x]$wH 2*H:NFAF!#0ZwFǖ;Pf^}iJh xPZV9Սr}`zS޸ƯK?Yܹr5[з[$EStM zc6;s ا@kmRd]93n`ܛZO~໾*ilvdӁn\5T/L.]@Wmerkuf`܃24GۿslG՗͋ gS2r1%*JV0 =$<Ґa@0Ey(ml$j!Rh,lL >'CR)A؝U#1`۵񎧾Ⱦ,AGyTӾ@;]0(ɤ({4  PW]BEZ+E[e."٨SBf*Qk#)6Gs2Q,hfҿn~3I`9FO' > &xa\FmN^1K(6"GS*0X1X#Ծs]UMG) n,r(EHQ8WkdORDƈ)5|ѱ /b@g, s/0*OyooQDCwɞ6X3% ,tNtWC8D<4=4wyJ_-tѰ;roth *d TuD̨#Z4/2R i~(fv&|}k뿟ї)iz9c|Y|^,yNSoCp?&&QXnF{nFW)=Wlmj'nn<'}MZ/~_w7to|oMW\n`v3@^xۗ.ВZq4yREnU9%<$P;-c69M>hXڤTDΦ 3@(1IœZa5*&[T¦Kz.琳5#Js #Y"+ %7;^o֝B1).>Srl{p<$ХAZ-Q8mS0uəҥk5A8QZ' 7A%2D$NiAh(:'05ͺ@]u3yVXBSvqs4Kd,Ww$Ybz JVr:w_v&E'|RoM{CtAtLc 𵲊 RLT"(4eA@ms-BTbuś-CON8*>* .V) uidlUaa3 1 wxCv` tdf懿< z|>_qf JE>d/ tZ\ Ɇ@ ¬2k8)bk|͔H Ҫ¦.Y)Oo'5֖H#v_9۝OޞB2If׷f3R-+ng 2. U8Xݬ)I h!yAa!>\/)_M~0Fd r9Gд2dnqlܕlqyW6v~rm7vZ1yQ*]! 䕵 3eTkބPb&:Y>v8rs>r :ϴr[smvpa^ȅAauaWyk}.8-W+ѿw?[*S~x?k|YQG|v;}`SM컜u>ߟ~g֊vMKiy{V291㬟|uַ{2Òӊ=e8y&XS.',>ˇ|X|OC=6}}}\~~-:SE[l2_-+\BQauMowOޏG7mW^1.F+`RxtN'<#trԭ_MBbv1. a<9#S]QYuu &{ 0oC oiKyB}`5װBe A#JDB/|2ȢcvT%@U%otUջg=7u֟+ܝiu*<ّR/@x\E;y" KG ic*]- %^pXv`vz@o앑1-s8R die:[vQ X@Hb`b)%%TWlwG"M(Z$sƓKRK#IdQiJmsicڄ}$1}i  `IF1wYZ*=>gTA} Ӛv5K7C;جĢ'|!ևbEo|)ct( /K ; PĞ}9@⤃^8HʑR;"eVL$Q f(G}4@RȠVD)L@WQGt)$ʄs4bg#c5G 迖bUϗ=+c*4Rc( Ni6)$#EQGQ0!S@VOe)AAY7Docpv|laΦra%$}ÒRǥ0 FUl1uW=z;]6 ˊDs\N)U 'PC5B`ts7HӁ3grǧRR /:|]r>X$yQ")9IE$s2JҌ;$fD00<ӑ'nX*Tz, CuߴAӚ'.ό%[eya3!u")%š:ѩe躥饖&!AHJŅ`v5$ȭfRŵ Rwىt+.`a# _3)jlZ Btq%ܥQjX0ID|ްSy7JMS2+^:<+%DPU4h_9u`v 2^4}.1LpwMmָT(O16.|{<>V]gv 3/c3w{_!bhVEfM8_/ґ0?9Z߽ K3(bU$$ctC#L9+)(ԝv弘"w{:W ;f`! H ˥CRp5A td}USb0J|0~Ȕڟ԰ylʅ,Q_{qM"iӅj/}spLqN(ъ<@!+ogJʗn]UMsו6Œ /f Mw`b)s" *N@ .u:D+zO-KvʰR}ٯ?]:s#sb*߸(@@Uj-`Wd="jNb(S{]LM! f[;"ӎXrmxEfN,Tg{gePzk?'DV9\teRbrAs9%6=( Y/xrΦ5qK92V9T 41-G+O605eXloL’g2DO/V1E4N1{qJ7DdI4 uJ() ՄʗA5E/)Jpyk*'H( \*"U+[CW fNWRκztŵf/[9`ў`ZCW*7HWaLQ*~s)Jpi ]%Z:]lGWo⌨Up9n ]%T:]%tt銧|bFQ..ոZRd"Y_a$',C D@OۯeѴ*708ox7|!J[GI5n*JBKQFۣ= ")em ]%XZNJ[+F9!m'~'f?5tЪ?J( 1"X ]%ӧ+@)Pw`J0̔j]`N۳Lpn ]%\:]%wtJ"Um3ۨdc3)o—%೔"Lqq0拶|\>\bH.9CPFPE=Ύh,cR*N2xߕ47?.xͲ=geKֿ?KD{nL6R?D hF <ȑ-t~|se~NwSFձ=/,ۻ_+œ};35=y 9h$v+ԡ\nGa(ϽTԑaZ򘒉*8rpA(* BYwD-Y`>> x5j"%+ >ه)|/n; (3hPfߋXl9} .ۏgNSpՃ/Ū6ָk:~Yw[|3ƌX^@1AE\F2<Ŝ)&QQEu9|tJ>u`6?}=v8mwoqvG|S|jy_[( 5}9!#:yDk0u* |= ôĠ(#< #$]ո#^8V;<Dh#.H!CAZ]asEeX(%,Yydc`hJJJ8LNJ7҂$# Ccp&c 魗&7K]?=US&[/(]Ly]F 7gxI߇+' Ez-9K$s9i<0Vv*N`EXG<h2ly6!G0e1%D#GT , ׄx¹XNPw$sݸA3 y"DWfzHC< @ l4TSDCݏg V5y۳~A7OUbÜhpHS2`=@6d/5Aiׅ/bn+ccu4?'ar>=NG^c],kGۨ')i h 4oPm'Fs RlV~5q4 M@ p>iU *-Ɨ/QyIG9V:K8;{F =-T\B)?FԂg3 ,Zؑ&MMUbIfUeT[9) oe "۞fI|!euI? SjEz O3 ˃z.}JZڥvYŒnŸˠ ^$k] J!$A̪[~m)ء"&$rmclSeٛBkʋ4OlZ|ƣ\vődSA$)qX׫2aT)/ V[ ܑO`WP8}U,mZ<˙lΩPbS5Yn/X?ߝQU H~T&vK\|M;{u],?{׶Fd3h~ٍ \e)R&DE4(Z D}GRmV..U5e7ٳL{^D. ۿھӾ?FԵujQ__TwDˉO?W_pX36o_oZmSb7Z9.0x4l?i=LvOQuߺeNOnewax1.,:h_j}wj6rOQG۫@)Χ,L,Mou}Hg&6y=Vح/onϮowli;ߊV!Zprezs %ĘAn1L%h:ؚ9]Ru seN%+n]RMvGaJ6ʅh 6g3\!9d֬,(ӕ%s2NL Y7c^W50+Zfqf'_ne;tz}CoNωWC\8x[5ds瑣Ti{Cz^QnU-0:0nf~DkJ 锌*$!ZaŸ%xU)1.C)1%YoBgҞ"[eaQw [|V(̱fj [/x 'gU|3ۻ՛>].ER]TekHBI:xuOhY+Ih'&5:4P#֨'(tk5;O=췌=uz`<~whHc+.10J(9mϝ+E9sgɲ?j<8R4P!a4 CPhɦhct4Ҧ8)sH 'JP]cP횔v@(i3qNt1P@]9܊yVI?/nqw˛ܩ%HE]t_/w}R7(lM"Xlʎ<9E3{\!<Ԩ ijRE6ޖ{KD3&99dysc!SBQGBAx4aV ꍓeP>EZ,t4te @{ǵͨUIvQ@kJD)'%& yc6@N:iO=i]1MJn,`/ 2M^'3wI c#k钑 ţ+4SsRlu Sl׿{p }@xf ̽R`Nqy`n3(4y,oNsm~VgPh:v>q:]2eABH1Hɵ1?(Rq[zvuNs^>ҥk@V>g*?AnJމK2RhZMs &~&L +蘌 'l$KNyCMjULJ!Q=>h 94>HJ)Qz< c@n^ 芨;6mۚu;P\$~֢D0JF3$V*AScl',|!#26elUfq-ƶЎ℩T˻b\]wk'uߺu_l-rM&Ksp>;QУU\{ JXP'2'o >7 3$ A*Yͦ26.BGW$T&Sb.Tg}c{Lw70"ҏ3-m-\E|/)#]Q'-bV--䭩H%)J@KeL^=Pt@q֠+ ,RLH!"d*c$)krcL|4ԁb)0Xg3-9.vQvqy&iRAsC/DБc4%&R/̷Ivvq(vkUc{γ0aGj߬kh֫4>W$1I-<_Jhv>c=>`+}SV2ٳe1$WY9;^( ;q{og ;%K|"ϒDAR"kͥ39[/&xuƜ@jTKe$n{ ". ɻ&xiY@8^%5gﳱfٶgn_>|R0-:sI^zuN\n^uJ{:0's&:VharALaI>/^94^ȋ`9UT`~mZ'$up)!+A*+@*Wy$E^ +) s>V a#wSoNs(kDYr*Qt=]vDZ\!R*6oԸk[!5:5tRcR>CRcgvҼ[q3 LMf; Q1 p,xY:Rad}LB.("u JzLkK6 %4zL9pA*" -mk(*TJω-ֹ(@Gɣ!b,2K$UIZs<9,4۱Cq5۟ޱtU=v5HKׇS;HrwG`/^?"E R$EfP$:T;w/ͰS(LY.ppBOfiT;i4n?z#Is 1LXtOwhS==|Pkg!%O H *cPJgr r W5Eo3qN$[Cل}+gmne2UIђoSA{@i[iՇ=òӣE˓JB@M4`Vޫ1"DIh h=T.?5Đy+HK1|K,ZѲMnkTO5YZU`QZX@F aiFdO! `gB B^*YU^2͆5YkOS|^'eϊ' A12xtJ p©hF4tD2+C(Q IDp&r)  R֍D:gdШ6C7I*,қ$$ $ _J \) Pe8Md_\~}IN5@ɯ] L]N<+ٻFW%^\YɇkeOIvA{q%R"džaIΜ:}zru9EeIVvu4Ћ.~>\Pܫ0{_V0M0ku 7ŧzǹ,~GYWLg5Ls\6lgz7&Ի(Ձ<\\VL,oqEq3EK\{Lbr'<_m.|‹~&#tmr 73q9]-.5n\Y"ZMDL+b9N*G3=\5QTWTU3iEsm#Gqf O ԀAΖ}Lh|rfwC{+ۇF7\,o@bD>nR6{Zpe&e~E{g0wy׿R|ƁyS*5 mOS9{0{ Q$n,VAT#}q#TKq#h(#)+cxGz'rɂ>tV $C(( Dsͼ ցԖk 1dJIt'bHt)2Z ӂoh4_')g4pgHb)na©3LLB2XP)!S|8 v< ؔdjV/TT뜁G)Vϑ%ə(4ԇ`^yvzzUDDs1h.(3:;-@u4v^~ucRrV88\XNI{-nB錂ekec'cpU~-mrrX.k9P ~lN!t:vV۱J#Nñ=c7TPz(%ĥumC)崎v2 K) ɸ3^Is6=Y JUY جHhcTY*m$w( _"S<2JR=*W[T;p78LUtj8~/ڱ4լc)xh3)Dr&It"XU2cx)w$^ADϑD$yiId^LiudN-+-Qi+$𿠽jI 4 m|OB$,KI6e*ƚ@|aU d$~lq]fy q](Gᗯ%8మV!=)a*1Ti)PEJmE&T+Up}ENl eǬܑAPvf4g*GԪ`G7A0-d)s\Q[.}90O0Q FƹLlz68[:WW|s#j8odQ#'^)(QL_@݁ŘNFфHLGclXTlRvwmPwCkTwt)0[{ai3Q9| -/5^ҧJ8GFTk%Nhj# ;`nFam2 F'r8µdV>zjd&$렌%,K!PIfsd2b49'C<>n "&B2I#pVQhʈ[L ^\Z̑f)K(rF,e&LM euI @ĉf? Ζ~vz>blHLuvFWSvkI&1i`3 ,:yԋtbsނlF)ƨb b:U2Vߊ`#it+ T6TηHuFjn`lB 9 =pTR<8e4U\Q{,XYȾޡk}7l\RVK^C>J%$0d&)D:TSnSJݔJ5ְST(!m3TC .7fz՟\ {1q_SNx4ɛA@G81?y8HwfC'~ZT4c6Ӄqu>snה S~ ,n|_L>]F,isu J29-ozrd STwI)GC~z.{ocxv#ʝcE<@W}r5m]8.BR7޵ul_.2tUQdtGqB YIѴCwCH6OWUZ{z2{wsā;;#ת rLQN|BݝuO_8m(7m7zlkhY֞m eK=lBҕ]>Ǻ2BW6 (K:C 䶤 Е͛+ezue()^ J%x΃n ܸ"Uz2:K hLʛ+Vv{E/tu>tUsU-EW7V h{_|*irfVK/(=ϑScLtqu!hC4 ?~0[iCV?`(eFi: )VW6* m\ p ]m==wщ?9]8*v@4[DX=nSGqA7DWvcЕMq+tDnte(.tutތBWvo{zGTU%j;^hdnl#-VEC!3bpQ mk(re]<ʷ(s[R6CWWy+teh# eJ:C,aCtfx9Ёh_;]$:C4 Еͨ+CV?ae(e*dU6DWLo vZӕBWgHW1U!Ydpf m^E$:CJg.)EbNRn%U {b+Q6|1lztrzUy`^qSO1T{bC)] ]=(!fR ]Zuk+CBWHWXY-O Ɨ-%%цɶ7n&ڜ֞m=.d[ve!`/0'@QNWrmQ.tMJ< ѕNq3tVv2.tut!o8 ]\Еի+C.tut|$+u3t ]Zӕ ]#]E)2i;C7of͠dU"a dcmQW73veh5 e %]ib+ތyyAメTfǗ;St7Di;S WVhО꼫P?M#;vAUO}apUahBֶř+SsǴ2)mr++CΑ{3vн+\.ixAAa,̗-5;+\H7Q T n̐=qDLr 8 T7BWgIW)2'wx+tehiVRΐT-ѕ1| ]Z]4Q.tut̛+z2NW2ΐ'~Ct#G+˛1vP: ݜo#NM~xӁQj/?}b?!~E.nSYH(}SIg/=o/;R[y_͇7x;}~ {[v_E=]1wO&q g7|§#>91eN)p ʿ5']O|}ܧqpwFS<+?#?7z{j _H~Ep6{?Z:P7W͟3S~1?}B#>C|PO?b{K]:ʻۛ^w?h5Bfj4C^w>,]r- |aR7޽/D0Cowhߎ|?/; uvuWJY|s.zfd{[]EzL8eɉ\OkTKkB_+qpکux{PS?ΧJBw짅29%:e9e{(tUFo(sT-$he D(%jyVDPHsj=.`m4x>j^l'q=PM%S?s=cXc@P"X 8jz'!Cag٥W?3f'W"eohbʼ6,`cX[Sƞ|cpZ K.XVIIx0HPRPyU 5tJ|Xɠ,vzeߦ@8Fߚshu ,:^ Lqja3{Y{mIFe0\Ad4gS21 {:}CESCԒ>{ JJRf֖s;QO-`Mڦvv<H*C8i@&E2-4؄ 'cYRpPR{B?Bik*3ԹCsPA2Y9A@&>;dWf;.͐jPo]iڑ?T@6/[M@CL !A,ʈLhD0em䚯)IV @PS޶ƱUH1H]ʌF JHG6Y=TO4+UJ|[b:&Y,2bB$&10eJ#FZ]iB,d3S_ݢŊte1cE5#f{!pBFZS LkW޽ow6+o^9sSCoQy̘*.fTdQɒ8Kt>SaZ*3d`6ayR@ qF/9*XX芸A`FR|$QL&rZd^je%U102Xh T5T Y/ ꖛU<oĝ 1XxT% 95 hd*!7[A; 2d~=~'pEcͧYuJφXk tdD"!;KKͺ\>&a` ^sM A _I`0ePExf -RhBr+Ιv C*bC - VLGՌZ-V稁'YEyB"4g0A^B;e69*I@"5g5\ u"N7HR`<(UTt>VqVM3-wPk4+hz<Qk9LɃZ1ڀJw^mꭻޫ."\ ivw@vH@}AώO޲ts61}tZB]£Gp޽܍_~s9v,Bg0FTtө%xCm䈰nL =4{ ZUY:JZ\kIlxY"4jx31G⪠/QeFrp)a!/G\;9P./=tP+LtvJdVsC; C*h@+|-3@z*\qa56M f+ !/}s^Ah;)RQySlfuIJ "E #$KQfFr:.gu @:nUj]Ўٍ6znulRᱎ %ߒas&#"1ۀaTC]s-/l`<14ɆPMiWAHcvܕQxV8mP jxw(;f@F\Y~BS+/z|dNGZ 'aiPj⎖M3i 6jJ\UKS6`jCv.fjRH*82ub*H`u/ 6ˆ;)D0B!Cv-׵RvqvYi'#.]sr ޚ_`N_ ʱ3Nf]έcIzw ,%㠯6aHJ8}Ѷ(QȜ1 F6kv7W7ԳW[^`G&F?MW.*n4f>hdR3?n6{?ZŦnU}ьS.O>%zf6m\~-һzquyɅ`_:͗Ϳմdm˧cFjWo %uѱOy91c{a/~|Bk+TDC8(H,'CF5O$P % T$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D/ C"0z8$P `H dI t8g}$e8 D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@*1$ g!P\)CFs!I 0Fh""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ":[ȱ{5 F1ȱu4gJ0"ΐr:"H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""|HR1x} +ތ^6\NsZGqv3]>iw:.q@gs8k`%t%tJ#p/^;S¿uHmތī>z(7uǝ٤IVwª#br8_|=ϖKSK8wt;z+/ ۇbG7]L LWqa0'͵۽?iJ4L^^5tI'6WO\=MZ 1WO~+s%\\Xb@Tᘫ*QC1WUZ'n\rR`f9Q/k}fczVy84]Zr< q4@KxuԾY{_&rfR%* ޼}˜W7fv4%Wm=ŧij׼-ޏzt׫6O.YU\FEM"ގfw4V3ۢV̶UZ>V)w,Ͷ3J)-ZF\Uqa0 uB\U)#suJj ^2W(b(J+MURwuJ= s;5*C1WUZ}7W(%p su(Ԑ+X37sUŕl(JuUr'++6 s 6`+PUJi534WNs ]Ui_:J,s4WX1"FB^6XtF9Ak4 lp*Ld>]JDfzfZigʧ *$c'6WOVӘI){f̕"sulsÍ+تᘫ**}7W(c9+FIcv:[,qqbDg_=f'vV̍vYί'Y%zp3A=ʇ52mG wϯ@ 4@-@)ehӀOG P>g1Գ/۳Lʶ!YcMY4JԀժ.yNv:Y|}v^k˺%Ѝ)4$$f)`MC]v۪~2v vyf0ޔ_/?Z_MAg<8mk+`*!gIc >;Ȃ;8Yjկ_j}g4dWNOχwe(/h!d;o>Cw9q}qnGm~ݝķOs/lgED=:Kh)ɒm o6@܂>eˣ,l6óٹSJk%ڭ_ZW|ġ~k8ȬFkw~f:ũ|`0W%`ogoݶ˽7 ŹwU4 |>{=||1qE(=,­_Ӽ\v5f㦓us&SW#_%t#)[?^vl2fbx޲ Rr&@*&A'AY6fЩe)ʠ?hʦ8`d 2|lqX&wR E*3'/Y6.+"B 1 -b⬤cG|9eHN8L8YTz6O;OE;GʥqIXV 8߂&Ȁeev\l,\ߨv ʄ(ZsB$^S28*x]Rg`1j / XR]/Au%΁.iJ4r Mc2QrkKA;9 I&|"rRWt6wF(lL>uClF̼F#wG^7Dp(BFAZ%> /)֫v`4,XtLۨc)FsAEC+qQm8O6-a:" &둎s9y~е>ok޿D'ڽ%v||Ҍ_htb.Kz]YsRL7!;r:[Ui|NF4o]Q*\*ǬiW)l:tL:-=:,jktU tNEBrPElz2%ř(Ř:^L=jJz>mQݲ>߰Ӻ_MbdyC9QG'GNIJvnN/w=z⧹ 5F,ߺ$3L!5 G:HlJ 奖F.Y7Q>}>ꉽІQt;z8D"Q ",2I#^ 1+.`[Y#U\GǗ8_nnZ>Pu2u-yA)Jv)P^k8I<^ǜ| Nj٩ibJOLaШHRZkSDٽ>v]3qW׵^ٶ3jǎOYo/E,OoԈ6u/bbЪ{w [[(v=QI]*@1u#-FEYl[#}Q@ 7QE/G*E#$PaT(ml6%'VȤBtPbwI[k70<*ÌV2 'M0+xܫtIJ;m*':8#X]VEdJM/a0-4(hmt^Y,iMW&G]mi.MSL@24&HG!Y,</;{ [-'TN /"{ا 4L;(|T^93 KÜ L,'gx]U2K%Ek;V!^0]\>"~շK4E *K o#*6F됓NTk@V wV'Kҋ1t&Yq6}6AGB8T9EͫsxȳyR.$t_H-O]"8J~#~-gXRkP.SUd->s:fJu.+RC]UEnF͎_ Rkkְme\˝D |Rty ߬\,/Yv/YF0u \"0Klw1(kU.*k% O|{r$qvRN):DjǛkJ6Z3D/ċVzG)/66l~9Rَ̰_HX,&0lCQ u`Ftxow-3G+|hT:DoLL;TF{}m3X>C+F065:T}U1}yn g۲^jvim ]bscϏnGkǜ\bBF+)V9L>E%!mp&ְ㗑䝛[1G80 mfc m¼ c5l]݌^3ĵçfٛS)-2f> x"FE%`p*Y!]_;iZ4!mUVBN;pX[!ĸ|B)୍t6:Z9$7oP v;hKVD|]\{etI]J{J6J2߀RѼIB^K5n)/ ]AH=n[lYVϲmav=!]hHb37bZܢ6zѸXݨMa_5"_j_5 E4Q7Ahxۯ}(B]NՓtߗ`Ad.c)\qZJu1No%EtYFS5Sr)i )cp[.e"-f9riBWN8Α/B%9>)12V"ͼ#(!Sck 06wd|XX` `V{id7 br"𐤧IskGd&T9}j_ %ogytE7 ʴ;#}y$UE=rM+IB 2fˀCDTFI;&Q,I{fchi\`Q"Makp+2TmdadUaa58 Me, ^b]`Pq6v7A,G8qq4Ϳs&rBȍg[,)`9rRM):DRU4fHsQ0^( ^ˈF1c6IsH4:OqżԮfڶGxTK=B?ȔHƴ<,1#1ss -Dr4JKfjw`y[cWI6y'CcE#}rWMz Y}e;n|ǓؙOҥ~c%Ȁ%*s(30au e(Ð"Ӝz*5m䦼w}*M:s/@Eg}N*h'J2kZ~6gn_Q8.C k9w$Z^2{56Yl$1=_}#*E&ڢ1 )WcgMslSa ];)MZF+}Tλ/h Vtu<ͦ:ܙXC%%d c!WIF- d ZsYiۻi@K "HRDHƖoCt@%Pi]ѲX;L[M#wլR3TrUdZR/t$Nv]r;3>&YTt<-ȣ7JQANkMdqre9Xz!}٤d4pZi [0z} ;64uٱ@$2`FəJ r4#44PÍ.B֠ Mno6Y'1/{#cOzϥY0ɘ ttoobq藦o45HyFi g FkSːKe]BOV9I*JE MDH\rtdYF hsLrHB_3]B\I]vFVqY dzZK(7͸y&;ͅbgӧ.F(v+.GM˄Q;p8Hsl)͢ 1W|/]@6 ~F&ͪmL`W.0Gi}Z+TZ?5q^M&$44M=/mޅO?m;,BgӨ5?h)fg~3:DP7{!D h0l(u)׿*ꗢR/^wo0qDJ\oV$+n+A7&8=A !\,lNUfylhF3!GsKޥ[a`hꖝkuaiiJm?4F+ ]k =߻vhhڴ|4+~yeǕa7a2Ew3kqGEhب`\lM$Gdo䇜nw[c%H#-QԡX<*(|:S5lzݼԩ{5@w y5Fσ< 0q(IeL}+JKM~ѮD+|_/yϷ)0|\~7~OTi{2fō>g4*20 *2A(Z q빃=.^ 9)nǶMZh<+ w_0Bd,̚RNnu`;2<&SJ'%c0~s4hn9]qF)ԮvV6|h #xpb NK ruՃWӻo?PHRCI(iƙK%L 'zMw E{EݲG}JPp$N,ǂ)/H^³ GzQ'|wAUNZB(vry]!0.Lsq$+Ɇ ;t>&QBH@BAm&^9rdahA4S*+ ]ndεoaNykJԆ_"O:28PdikL^m[W>:T*_ LS5=B>{p`xeWϗn;,w7y;+S`D̿he/g7xHl-7ϋvq~(]Ea bRdr@oI㬜3,𗿗E.^̕/P p{˂+=4BEe]#/?ˋE+[] S^ s[P]*~Y~'U7e Ti|lw=lQs0oq76o#[j6&.(MU- ߪ`V|MI#< ?޽zGӪ ~X=CSjz8ԄJ+[ma1JuI\JT,$ hm'i<#3$ik&"Zʋp5m+LL;vJnMw4lĉ\FiZnOCWP*. JvtkcӚn6\e'7vP$gC Z.q̓%ac1hd4ceL"ҦgLaOι\SYΫYae^$[mQXoToX^,04PQ"'6dfӂIe:ݙZhhf[u,-me=c7۞lˬ`RDDW3 ]!\qh m;]JADGWgHW* KB+Ki,thh;]!J;:C x+kl,thzKK QN:GG͡/]`+KŇ+@u P 9ҕ& DDWXS ]pѮ-m+D)::G2m6& cmPtpU4+ִ]#])j6mE)gxBf_סMQo.4BECWGh%!miDɺpoՖ]. ƞ^lSǛ5kOU6@a' hlj@W]1\FCWWX vΈ`i3Od'Wˍ:@JB"o>T>sV[˦܄H~,ۯ=| WoYޠ1kL&5(Z ^KgYZRO"͍aN˄{ubY)mARy gUj/R<}>g0YIzvfXvTHLj}4*>0<$&1ED<6 $R"n JֹTFj]`#i4tpi/T3+g9 ]!x<ֶ^mttu\$HhXN++Y_Gʴ]!])n=^tOd ms+-:"c "\J FhAD;0MDtg•6Bl[CW)ڻRJSrm&&#hhhў*M3 2vhZolN"0? .;1]5C+NDWP-n@W]Jq{+9խ+DiYGWHW]ќ'uR6,vDtˤV<&[l pn϶LҶtDl{ٖ+[B8@RBWTI.Lutut%P\GDWX ]!#,l"m+DٶBWRXDt"Bʶ1]!])"+]!`]\CX,th;$vBBwtutŭ0FDWR ]!\ ]!疆 QN:G:&c+Ƣ+++DY Qҕ_ILtt(X JvB3ϑ4&fGCWVʶԼ3+C@FCW7cjvΑElp5)DQm)Q,Qm7u^Wf}_?p'Sɋ]ƳwWV5jZյEߌGY麼> /֧%I؞T?ߋ?ލgay7#_rxTm89C `DQZe]ρ L&So磮괏/_}[VC6,tԻ|Hp@`0MJgPt\"xB^d^9b Dk+&!jߦw⒤0(_'׎I&*˹(cs g2eP>0i5̈ז/Z{b8(/b:./7_T\^y,;Gg!̟ {$Qa f7_7nzk&ɲk_bq;qm5zmo-82TsmF eE0=f:3]F̓y:}"3chns&ArB/6YnQ*3xo`42>+:9nj& +j&=zabKs@ U\r ,㐃9R7Z4h7=h*z4~,C-\pI鲞TG-Վ_` bPXd+ B 0x- S*Jٜ *32SL%l-]jn0@/!>'e.~ҩa8O\.# ^}E 4Tt=Nd_ڳZ+;XfӗL`R f~K~}* _4s>lplW"ވIxGuQtc ,FD%DIJh9m , 92H-Y{YaE  Ճ [xuoyɶu7zA|(ejp6,\zqr1_IwC2b! AҚ$VYxftQP) *ɭ5~!$d-̖][s+}JR 6li6Jq@CbL ocH7ӠEɣK@LZdF}nR 8̊ӚF78-]vhmzw_r]7~;(Rѝ>Sͯx<rHV-yq]wwʭt)'ю/ͅ0kMS-!fl\.VkTdIרm*;ٽ3}%f$_;ܫw ,ۆ.o;_|Ynz]Fg8QڦK5L3X,>cwc/88ȧ/-8'vd99ܷ0 #S̀2DDEH:h!ۘ*y1hCI[HL(Z, M&bH@pKD2"s<@l_/ew}u7/mq;۞F15eqӷYy.)=fwb\= "2҄Yr 2Q'Վ@)-V3㎑Nܽ%Sfnz<;1y?+s+etDǜ8/[;Y"s\fs pD(T`x`@6L sF.A1$":M\;>庹lܪ,MLIY,TUNxZ)4F \0E)2(QTjiHa@@߹]->~ߎ|v;:󈓒\Ϯ<⼽PUE5zeKnc {+bNacE1Ob4MܧQ_6콽O2_rw?Nҋ5*;kx]2=⣸ L޶i|!Mӿo*1s]!]dWЧto^_>Szԇ.jyyWGBmǿ/-]]W?P6;&<{]YU(W}Xh}]ٳ/_ Yor//@b)2Ft\} &`х}(jݥ"gJc}9]^M]N/ r_" y˺);hlq]mZ.ҲMe:^gӒwsf$t=XwDKRplst{@ˆg)NMOW5wwB u]Vŝα6б*\WrCFJ7sl$k}+FnmsB$)ñ?᱑R*c+0@ͯɋX`5R(rS!H gTo"+@ Rx졑;n/8bd2cq}+CcOm2WR?äK=Ѷ1ZFWSK @^MqƚF1Y0"Ɵ .-8L ;Ia gDpu.4-"@ R(K<$MY%W5"qX*Vː"REf)ỷC%glr\I5q  ٗ+B\ц# ʒ奜,OJ朣81+M}+<6mJy! >n8Ik^J"*X!59M)X\l9l}fNn)K=ˡwTMONf4]]P,W \t˰b [eJ3!4JtU!ʈXMÈpb3zIkմT\q8>I:!DtL BxБc49<[cI x\<⡬4<<(ȃob (#W~tsW=Jh]u&B?njdd V4* 4AKĐ|` [1;)_vne;x,c y~T 9Y2 `S l 'Jq%n=Z>qW<$ܡ7!u}e΀(e@;\~'/nV>l:WO4Ph o=ō5Cɬޫn:0}κ3z?WݬJb&ٴ^ n\iqFohf]??wIfVMS({PPM>K*?[$RKK.R7χ{{Wwc'BPȄJ^#S&OA 2%!LZhR\ۑ䍫k9婶}@a^(ZZfWmd]g,MT258йa{,>f {z\7 ^P*]tj=U*gckd< dzQ*FX^9fdue&Dr@B$ " Η V.H$[yڑ&93<饬?IaadLhkwЉؤ" X1<eGBSGe6XPG8\zzeKʣV1 LVi5]N2y1mS9,f!iR֠VuDqE ;˹%@ C@#u#_.BoNb=N/:brYj_XQAdFX@ʎ%!ڢJH-D-(4Sӏ2$!X pDMcRˉ!ih`Hu SQiKQSC$^*񬒎 9M(umZ݆VLۮ!QMe9Oegvb8JF16oFq͇X im7fL?, YUn\Onu`W|լo/{$UeO=/Iu}}IJr+xH=M^K?DGz̋/Y`;)ˌR,_=}ڹ:DPnY"Z q0-?~䟺?RS?Jtٻ֞8dW|-^iהּ~/wF7h4MAWkF1PT??9Y&*$Ci'TۊWͿ˃Gag>)sֽn=gJ|#mOY/ԫpmfkh!b (f#2EDbl|Ed[{@¦WcG^҄Q%W J!cQvMD0Lk?`/b@ZQ=~-y˃Er>C6m*(aqGB߾ZMpo-a%Q, M6M]z*]Zcj'(E<]qu{׋?~"T82]4:tN˜\4[[EmmA/ zȆϳE Bp"uZL+kt_&-} 6)oq)gKrkiIjʢ :TG~]U!遽< '^(pݶׇAw]q4+qCW՚T\vU*^]\JFKhJمsXIUHؚ8M(( wl~kWdr4y꫗w?Χm;ӟ'm9iCpM=I[HE.&ݽvgsa:7ؔF[\o en wH-(GU`uZmf[k69ENU[EU5fVt{p6.swW!q8~U-ƂJ&U,X[SRI+A.ۢ08U:`61t3(TTt2Z9?fLBs /JhnhI=YPpsB&4q ÷O6<PHY&Lm':..0ĉ񍯛Q5\;\ĩ@*7U{Pm$"?Gyo*-wu5p,XV ?D7ʟ.5ơc:A*+r\/(8>DZ*$MqLA*T A"\Y'՘U SbVcUkA %:X;` h9PUr |^IluRrޓL/Zk1^\(\hY2UNj˓O{ev?ZD!ch)GHJ6Gu0RVSy|g8Fv0N 9۴x^4TK'PU0޶*$dx"DOβ4ND[ksD/Ʋ3mx]G=GzÍ`b`-[uVfrl&bB՚A3Rjv;/Ǎ5B]~֧%]Mw> +w0(6ɨK! ߭P9Iv&RS1+CR oxݐ;;D%ʭUoi^cGZ*BW9S +6&Xf/2F.JjQf巋u'a|@ͫ5Ӽ)LtKB?\Nr;D],֠o 4wMSUK,y F1B9gWyBDPPy6z*PT S l-RRUM,j[K(l2 B/>y> >!}'W\-5/ G~i @aΫTN6ٷPy\߮ߝ!On:ڑ:bھ::>lړS8$M:MdZuU]K>YD[I"zkIDo&rF٩Wz MsYhLh`CztB,X yX` 5H= g74NJOZk`*5ZıPk(=57Ď&DW0tЕoo骡43]#]d̄f7T誡u0vj(jYSxgKS~ʼn0-#0?۟...4/NNсG-t[-tTh 3M74mySo,1\UnՆp6C ]mҨڙzZY?!֪ЕEES+A@J3]#]1t<>^`Ns<}w$CyPW{AzpvZֆlqy7cIe "HdˋKƅ+ oW|(ouz` `n[V#G>e2ʀB$Fr͟e X,w[ENU{]" kEL$l oӋPG)ǎjV]:aJ/0*7 h8jh_;$t%(=oJJUT誡ѿ Kr+ǔO2X#fuJj]27h0t&C }dCIsCSA9nt`ov]o3~RUd%tgv0!dJ:Е}ӕjuLWHWdzl6>#o1MӏkՍib#35ȣmZpmNFZ[l7hy>im 4OGipi2G*}ҹGC#H8jp ] Z0ҕy~% ]5xtP>Z nJ;jp ]5FJkgCrM5?{ƍ /'ޜqJYYU:e%jm"'~"QH\qHntk2Ԫ]*54W-1kk`j7`ЕVv~PH4WߠR\"2W@0F+C.fPEC%4WBFEl=$O2%L )8P  !%C2 jb cZ}3 Tx"-i7',d~{<ӡ͈RsМlNI{2Kxϲ*IqkÀN}:c-qT=+QOXw7(.G _e7÷ìhC9˫"gHo$Da;0bx #vNyI-ɗHd6,Mɦ'|X]arze矯MvFw̤f++{+IuyBK3oIm y\^?ǷsNdƀYÕY)𻁠}6?/Ga)=(pS267ߙfr< ;D#5W5r{O8Qk"q1]7o܇\O!9'0n UqqLUނOfp}H Y}Ly|=ʃ0K]ISHy+f"wSh vc*O1Qd4a#z1ެzG]!kL8\{f" IBo&[Bf֍s|r{;ps̗3 ~ba|th [?}W:~, )|psen@`B2ZIJR0JT".TBq_Y~w64n?5bᗑ{_]9ަSG|9AԞ#D_dHM~[cK"Dht:L8qbSHO+RФ$$SJ%+fwT-lH DB)|oeʧ^c-ȀZ(2/_|g!*Lq|e-AlT kte@ЕAW]5Aj{_BA.JCS i;;S_JkP_c'Tŷl\҂﷖|Jǘ?.əPO*QoK0ZފdLF5F-g2sd)m4|5cNǒAcכ5f7?>渭h%Q-+^eti_gDՏ {;Mm9+V,2 / xʍ:]Sʀ7뷬j3嫶}@0_<u]n fvB*; v"09}F--oI.N{W5ApmZl&z]$e0đӎߙŗlfn;̥{HSoVQS]^U.8CH%%N_/8 2J sĊTUT4S:U%B뢠UQv(ܧV!ᘻ-&@]2#ug>#)h) FK.3gLRl/[[,[i,ciBL(+l`ƺ^̘<@f?!)սY8ˁb"P`dn"!3*brq7?k)g]ǬQOvBӀ0rT_ew vyմy>U hL,3̖hL)8 %hGעmfՎګ-o\7ʐ2d42;!Bv?ŮdtĬ?239~BV4bN ;l] >6c hĬw2bƬ~H=l#nhL5&<Y$<"'eq5ı~ձA#X$fuLg24fb%0r%ˁa;ݔшrSY)Q4@!bV[A@{4Y} ZwWn bqDUi^bHU4gKJxa"PjjF(󊣔 F55%iloK6[!rd?Mg5?*ӈ&M&Uchjg6w;՜Vy6:Ի;2*D2aHDJ XW-0^\~b[[w,#BfE!Y%KDo<]nsW/X#au5Eie ̕[Z- I!ڸy:ο0[SsXD0Qce|G27(ı`@~0޽ c5*_S=0K7V^+b /!V \nzfQ_0Ab7]iInn@%x0+h\ :֡ރt[nx鶗ϸE(%芃ȢX[(j %&b MB(ӹɸBrFL!,|ٿ7yW)= ׄ&(]b`E&M6/gwfY=+Vm0/| #jCR$6O\'5qa>ͧng}HNVlS{~Kw6_thK T `:~g7ж !`DƈD}L13gL5cBl+cc nk.{q q(RۧIL{v`%N,u^B&pH] A&5Մ_.ZU<ӡ%r~OۓUR0N(Hs:q]ESJȉ֡lI)g1 2Np2Q7Yb>r__4)&* N `pEUUsJʔVJTe3BIQ" pPU&;Nm!0 &jF׷԰ Vᨾu&B1Q{zO9C1^ʫJ{[.>k!t;2:& gM:ۥq.v(b 1rTU{0{o^|ɣO@>pځr4& 1qۘGy+yg1vt-#}V1AYcF6vowǓa@B/ijT拉Gw/+K2,\9,^5&{VYAKiRLLùN Fd+(Sh?6tz=^l2Y$yzmHׄvVJI_p?7=ܱ>Y x1=2Xr9P%ׂI^c":'0qsÁd$a60 \6>ɓޑo Z0&y#m't>'lpDJZ'3c+)5-\gl\4< >H"SyrKDq֏A$+Z Do[0R%E?fǶϰRZ0Hyo}܊|K}`HqmJkey.7F7F!FS>Mumh~$'pԂ!!3bBMJJJWɧQ!VǗj&<mٽ.=a"Ȳ"Xk%c˦5OfO"Ҭl'1)YA"p=JqlI FWHTQIIDIf1~~!1AٻFr$W }=c9 ncޙ`dfeU\HɶRRDɲU+ȏ"-z"Vji FѬ:/f?hkS:a,w"^ YLT&K޹EmJŵ(-8 _}4(Ś\ZA tZJج/^V<<+ ՘ӗ!"H<% &&<(D8Cj*hZŏd ZsRms4?8ƼMWwaR4Rm N`-~p! >mIʑR0I5}+!{Rhfhu#L_<{7 5:L+ ROmը1 ;g0[!} _ءa+ iaAGԍܭk!tikgA h(KSs7,Ji%ӑ ѱRaEOܽ@n$*I&G.a@'昕" Ĉ 0@P5t6SN D Vs"jE$ohi.cY{H4|&^v-)9R3deMTxF8%.h:-B_p|#,XediƋ8rɡVt#dʌ vuFBu"CsϑCfl0i:eSmj4:~BNA bK,HlVLX%JlMHBY5i@sn /rECIJ\곝y˿ռB*8!B23WjR2!0²N+!5>S6uMˍHvl~ɑ-(N$ϥ<1՟tE Iɔ\MN|4]ogT~RrMe,U^[Q?\ss'М /w znғrSNl: :SC JS4yu)Coh6p UG9>"snaנ|[1 NnN]S~3`G ||2yHj<ĈK,/fSԐo2D;S4` )ۦƔ02\I޽Ԝpba[7B0v%,=z7Mu+p.hy1LX0@<*j("5 ce7"0*SL4!5܏!y8!˩"$$6~)m 7/!A$^MAc(ɑ yeFb1Ls9q,6{YZ!*A7ͳvaQ4cNL= ֪eta l%x;$d*ST$rU+KB~33o}OhLJ}ͷ}X^{0n_n2M+>(x=5" X4– F۽K>Gɭ/_bT&w>*wQlT! ND8`EJQUo_Rrl9sV`+YuR,7cԐ _(kqtl豪sJZ'75A%*Lň Z \4n21h]@0[]z'(  {$NFPUQVsw|>vw>E <^0֚U6" q5*pU$~/l3^8s^7oI㱆ےRlwU2k,,]YYCѶBgN2^ cTH2cז:L3;@Bݴ$-ܨH,70(u_5~ fǽN)yF"z*Rʌ*|qEɜ6s +!E,.isߙ MA2^EÅbF bdT UWP=pϸ]BKUQrѸZ*J` UVh`UFۚpʬz弢g!'{?zpN%pe|\o;-hFCaUq?U9=c|x&նr%bev52X.MZHeI@xf?;WÓ&ɲbplUsWԔ1`{{^'gBfEϧMk:J(Ѭ Ք_q`K$ aJ&2BRWZapEMY=m{GT_ *r Uex16ˆ Ȣ I,qFYN}x,on¹TGΞ{-Κ֮Hu9ԕ™c?_7@׊؂Z!^ i%.RF{r675lJ;D9︂Dr]% k` P1Iƥ.n$MiS{iG}FN[8 )IΧRuYJ_gfϬ*A`x=c{=JZj|턳o;YaϘ D`(r{t.!'7Vx=ew~Uvn?)^^Gqr{pᎯ~N>v?}IY{DSޒ-|hf_$#B?{U,0<ВDY 3*]%5c7"qg2ռF݃-t}O_bsCWeL#ݹI,*}Oa%1 '_J^N>#tzͶge5;ׇhd̙yv}쓆s!k:w0?e-S$wy;O;7n3u2J%2*Li8\Qy_Se<^P׹&]tNW(Pmgiz@zi̓aQcꬮ_d_L-9"nu~̢-sҐ{U|{SP%C`|t#|*xzƸC2n4ܫZL?~^(9_ fL',aa]X'^My_1K `YqGūesg3`#LGWזHI{_AZ=yS!QY,ם1 qG.t ]X"[/y(8~3k K"}\{|Ϗ_㚩كnB w~e}!,s }鷆yS0%مW4E+ע<׀%^/ :g飞41;=HE<"wՃ!b%QGYl>ǨL{O|txU~tF =o*zy ~f-y;G}\ KvySwcy2F->D8ċY |LZ9c.;"%t Lya \cI˂z'b6Lx8)9RFP,R>E0-zūB0b`wdAH;ݤ"Dʄ";+eߪU3uL >i[#W М^oz3KqvEs^uMx>iqϜ}> vL="ǫ{v ˋx)4T^^:5aׇUquLo]D>J(teR9e, drNc5sW4A KNA>6!hQIe :$fm,az [ ox-tj~ L7RM&M_]в IWbe-(sϱdϿ˗i?5yQFc Wݵ~&*У!]? wOyZ aʿŐqi'7X 4.{/1YY/ŗqd\u4/9Rq8nً0VרXkM D~ Y0s/̂0wfǁ&A0 zե/Dvip%z.003)xAO;t{ Jc"fr?ݰZ E{ՁC;7JtT4vhϪeC|fCXga8oULGtLv{ tAj~ Nl9%EÇ Zy03F"0Hʪ.W޹ ,ns6Ӿu/۟/i`?8 W<ҽM#{MD۳0N(焬c4x崜kEy@Tɳ;"SlQWϐucyyx!3=) xDOAbbQ;ֶ3oNl{*`&z0V5L]h2X?*X",+ jTX-bkfsAw`2v䜗RHGZ=}{vwqY5KDoUR&BGud29Ѹ5MC:)k@ө8 pVn5Olx*,  gl9rnz|qd3'z(&.ˏqmBN Ui0qY0mz(kʯS݄bMgm3 OsbFbh&.ky4@L[1.M'j]U!C\״)D]}Y.8 ,-| a6sTpb8܄^,OL%6骉9!E*%B4jqEv1nV= m]{Cye%x^郷h83,/groMDƘ%C'K$OoԣHǹf? 8vBEuX~lJ`En- ʤtny}ƫР"TZG ^DUII`B4ڭAT.@\&v(Qb[%C 饷{{6ʩ={T)H7N5ԕg.uGxmn^34RdĔ "܊YEt8}yn,Q {T>M즇wS*ZI kxˢF±&cV5O t6h}NZnvwp0ےRl 7WKeAdڛVPu(M&AgxM|ٛt4 .r9WǑsE{Ƹ*ݹ_ +h҈SH ɐ xڰ|ceYU[*y.Sv"۾LHt5@ (R ;yZ\x}ڃ7Vlqu`܆MvHUq>P8*ZNt&Md|b:/gL>JYIsIyr \S$FҨ[m'SCֲ~jsRw)Ux+)c!|~NM]h~x~j&'ݠ&kR 84=YM']rv'_KY/C<NJB w l_U\:#Tq-^h->V ^zDh|&KǩZ,F Fl/!G .E;\~ΜEDӧpTL}e.PugɬКe2-`f+XA1=F~GEs?)t@3jM|I,r}1@ݼ.ӄS7; 1ihA텠LF>4 ̀`*Z.A?+I1ie}TzGVgow_MF7SQFvJ0b3-)V["j'VLO.[1`fm8iu}-P"G!UO!E ކvJj]`.|qF6Sɇ'#Nu4'4u;:0!h_}Ue"R'Vqhe +cuW$"3KY#{[h^Ej^^k(S`m\me )CD f_UJEhlK! {w|v3v~WDXw;P SB۞ 0lAq08CLJFҤfYvD 򓱆3v$Y[}q$"я=]y6иޞVfˋ6 (0Eq}c` %'$܎k>ZVӜg5> $Wx+u PcA֧Xg+"*q@`-Kpd|ߵB%>u泿2[ƭ,3+?F^!<G$1O;]$X ю5K<64x}1H\o /-|g|N@zĭ8@wŧ)>iD!G K)aZV9ې6'גJo+/e qsZH& 4S))FeE@C|7+[O|Fqa4uᵡ6P߇k%hWYpLg@cP w8(gjOg[a^2bZ`=#.btm՘Z -А.bˌTYo*ԾӸ%z_d ?V`r<gfaޘ~4ξ^_7ɪ&/7q[PL.'{V5唤rg`~4 7|L%7oS깔J:w}N+aT}泥SY8xյ46t= O(F<,& k,Co&}^Y%>"4 N-%uFu/HhB7 $AhaV[HFГJLIzT=Q,0F kb"=w3 BRڭ޶:*M,c#-BN64I>+Ԗ =K$].B<hjf#/ٸ/]̑M|wmq7p2Utg ljB0ak؈d|gWĢ 5a]=Vm*QpP^WßXלc""c u*[ LSl!7og_-|L>ywɔXE*e5 9)D6Յ,m3u k5ājs0|aߖ)?IJQ;r 9{;"6HRD!@!qFAqShl4^2B,C|0'8z[7bŠSоEmvJwiQc~$^"eD(6 n'P" 2浇S?v%x%{ZRjibhmņe/B߈)iw@;ӸFE.}^ԱNdnZ2W udљ|{yNyr.FyLz\d|B'OyJݔdf}m: W""NDzօȊH9V:tvy]r >WqU)*ǢHIbnk@6KQy)>"T!Q j{~| 7=߰bq sM9tA~\I#7Ol{ .l2Yxk?A(=(ZB*)cEǒaU I N*XUl}t$('fYE1F+9A6t+V;DWF#Mŋ"b wxɭ]W3Q1mظ76 mAoEx0Oڶ>;/3gH.|*Vezo&QGERE(H4bV6VaNDWނkBC.aQe"]fu=zX]e+v cu x^} wugj[S($Xr~bs/q=Bp+6}"-֩]ϹIְF Ztuj8H,s0=cR>>\>enMlj5Q,kZJKu,B'+H=X #i{E6v ?n/3Nc&{&ppddZdq3 &M2HVW~//54/RV.C~߸}&csS{o[?K1ȇJLZd"^>o`4X*0,_uճǿe"o6@RO2nV(v;TasDTnk`eZ*If?ԛGGs_R}|E"ފshK%A{Pp+9 D OΝ >౐~35t>Pǒ$Ǻ)p,/`A˺/[t/n5!(3:(G-4o%pP,|9 ry0/5c vQea!~; Ҫ;0<~)wRu,GHqn.E!eQZM&>fesac͇X6? ;=B ݵٽꀛȅP'`#¨r7} rdusCy;C\ŏR=X'ᨬImh!6\ԘEgsL SA;dݸ9XowfӤB vy|EHIxNcW(5X(MS` $ɑ!M )jh>VDɴրZBkAJ؂IHw03Q9CX,aPƻx8I9RZP+Qm5Bhn/Gw$t-6?ZLcQJ`$ ̊?gD{ 4޷اUcP̫jjNfja~%0wh&O3d^ A\$Hh)j.|a3υR@{:Ъ۽ F'*G@[4FR߅zkkhή)ӷ*p!c0aa L),.,(1 Q*A!^X\[HNR}MIf50 eHTBIhܩ{HN c_}Z\ƂYkL bZ i {ofy.h5w-նJUj/ ځn:s!@h%9-~gڇT ĵs [_PY ĤAb"5/,2.|q!TkaXTG"ƕ wqЉӡp?L^ j>_ԅvX`CUo?,n]yGއMp58ߨ}Ijtah7ԂT}8Wx@G4)TzÆ@Ԥ[ G~9 &#aY F|(%mXME!Q뤬Y(iqN$NGȑ%ltq7Js0^}&cWUSê uXGd*axo]d:ԣISSt?څ) g?O&w_ozKBi&\oe+T3֫\ @˽ON Ro$UE(^,nK,>~󻞽fzOW).G7]}|ǝ?Ӳdz͇dXN)A̠A/.L46*4 ,FK ISKXJVG^~QQbE>_Dq)hFT_ϡ˪EZFܧ(f2%P%N#eҮjAݨFZ)*#,Яy!z}5(& It'x  *lS RНFuю"/M"6hxxeW@r%)/2XBXwPHu YgxelfמP7)zhCqN#%Ll9RJ3DͥF?~7ULgQitT:,S:-"!a]uL0(6J{ rKXc fƨB '$My Nf by2zD4kH ]4NW. dk±,qubM b&d3ћixyΟHs9 L&Jbcbʁ)Z&N&92zs0~/kR<32F!WGl5M[H UP:@1yÊ|shep`fI<.` &ӱ[1 gB٠BNa>M¦:?v&豈%qd1&x6ՀRH$JEr#PeㅵF.ڝ;ν̋QI^`c'HdH~sŴ}9 qkY*~_&PLm;u v:ê?a9dX70 ~7/~ۼŊACנ?ߖW_GA8mt:e%_!Nz캌V8;KNuZǔzG?3%˝;=4MTtj g_륋z w$\^ܛпSoz˷S/Pj]uP s\em %_nyX.h{j/A2:ioP)[ K)UG *_ǣnJKǹԉr-PTA0 f˻Cy|ᥘKv†j2ϢT NR_c 2C-KѤp9TwNRJόVP{VvM]NQv?KeOfHF@*4 Io[`ϋ] \ F / ]tJ?(0Yx!f5[#o, 6-?f 6;:#L$ -JZD4NEo$gD,\)xh;"DNc\Ȼ*"tk42-gIu~ JF*<}6! `Vl9T]f1Q HMPtlz:Ab fGwnNjNmjR`D pKHJ1șD՘k\3Kztϓ3_pxUw|6ೲˡ$H:)ƄyZh%?8wV2la2~fYu5oW5zتg9G, V n/qgܺl3 <<~KA$i޽5"*5GFU<'+E]h4>t :c%R ^=\pV!MEP;3@WGDW6S%Zuubhmy2>eGAW E1;rǓ1M-]u`T (5Hs&~g96d{3Wc%n5y]ApAטg8I? .@ݎ\.qTrо)| 64!Cy ف[B607HR.Cl׊t:2Ze)DCV|\[֣`w߭LPw~%ٍ[-vjpP\|$>81'2M65_C``B.p " ? q$.q5z32Løոq1A FT \yn;cfFepW#?=Ab?J%RƁ &%Cecw1"l'Fv90% ͎*= lT̅NQPtA˵/89,';7'{]DL3V|@_#-$ut"mcy$Uc?il1 vczh(ؐeҌw4+*eyW3^ $[ ݨ*[]SϤ[T .v5Tu 5jT620_**k% y'YHجk ĈI8Vnof6i,p-ֱ`o1勄] nzhs b.}Y$=SWi22%~Zb"lI1K?΅Y !0LN1+ͧknsEv/I?–i̿zk7(Ӟ]H2YOB7~ZZx]'r1)G/һ_(Q4ΒDdWGEʔ  :LZPW! lFJOWʏخRg-!~=7b= 퓎s 8¡9`@Qi P='Wy8X&k«b|x p~r¡M&/n(f7CQ)Z͖?3)ZWc>wk+x jiqWX6F+w: QrTDUqB!\ܱɖC-L$`$p#OAh' YSIVU]ZLjT Dr_:ƐgX]#9::"hP0n> 8&/$8u y-emtetÐa2滛j-KQ>pc֔ҔRv|#2+] 1"С~PM,*#EsHmFf`mh{}(P1@ @QI0[|ȑf:ng*)l4p2!0 &_uhjKZ i^1kgo+Z[_%?tH̑J2>9M{7Z *& GF;8mz|[,a^cGC@'XXLv33QkBʁW5z ;bG('uѝųDŐVc;xp]idwq%sE2RMhS%rd|gVӁXlSM9g]ʑцGJ:՜I:,j9I:ǨԱ(q/LԖ?#,;p.Q[fmQ ыZE&󤆭=t6v%Z U\wѓOELb(T5pW|wCgFN=2N.Z銳ODŽ2RzWUR%p>kC_ d%JB#ǻv8W#=QWTv~e1+^c}wiF=\#c|ӗ#uֵ3g[fu9f=}6Rs?mqfd&~(>Mw< BG%'O~(ߒ7\5RMܥWеۓ?5zw>;L_Smƅ8c4K^J1 ОW!>{◌#V=guޟW%c-FZ 2Zr6q-BJ! = Џ6˳g ؁ݳ!z uj\j+J$Kjח%c|9;~N_%_%wN]1|}K>p1X"f.X0q^";ˊ8 DW@+EƟ3 0?^?|'/_89'?^KPVcGmeX7떣C$=52>==0D*!g/L&xݣ%.cu!{VKU$ Zr8XZXݞnއC\t{VY}Ѻl/2EkDjAQdR}A@Hky"y"w ]y"~e"'rDn#玟贅>(Iny"< }5~5:l|?D"t$+I\{y H}5o ͨՍD(XwHϋC>kh6FFm0;.#Ԑ=0&远J~<=;t$;Ϥ\:s6ÃF?~#{pg8tH&>VUGF=۬qɕIwc䨖ڲ(O^HDTkK>ud+%ZB\uu0'_c6?|;O݃!dKzti<vVml4rz6晓eзONŃvF;6zxjO%l9'PȪv䔣ZbrK yߊ@))%sMNZq֎-Ҏ; lدB`ga|'4aӜ?*";yO߀O?R '%ضU%cYVDQM[ b`ώ%qFz >2o?x~ntKb {F; ۣVE.㞕A`3wþoo'#ɾ썾To,! ٲۛY}SI~f{ت%cl,}1`>pCQ{YqQfJ]{Z_\8J:_>CK ,91v.Ԕr͖"P>k6~9K5\S{[2,iƦP 0ϋ{t1|#rlLu&k j~'kD5%& &.sn4*5嗀4\mrX}BeLjzZ\s![,"ldyE)ՌV3Z͈k5qۈti$"FڝHvơmk1s9ɓd%&4 Ɉb3O3=}5V :* B2¼p.*V CKߓhqܮt[ԯ zu٠3X7}TEѻhNyˣRj%< D<{JIYSJm-CE'^dJ @19kU%+a&tUgV-T]UbE]4]Cv铳ܞboO#,rV=(P?kQrp>)SKՆԿqM}j`+EaE Nf9_.2uʅWBkL-\/kW=4Q(ZzMyaM}Z*YIOP>IՇɠrĶͲAI N*œWߛavH2\ru(=HE^>$|UReY3)`?ammlA ]kf񾲒+^btWT4\!sP[}}YY/y^dȦz:~HyPVO 4)*Nq+U, ߠd+wc'TZQk]~JMy@ d@?LDw=*ʁM1pG%$%=Zur7h&M^t+KSlSS UIuV:z M &}:EFlUj iy;-䀺#yqe}+Mu"& J)!%ѭ"?jge?:jEbuUEЬi0P!Bl:8]̱K1Yy]X)D_BudqTb:uɩbU2XԫcdJ޻b-ǴD~"]j 5ul"K5dT5̮AREʖ)5 5!QVg "sxJ;.x^R4H4#EӌM|&5.첰 b 7Q7k2dKV>8Q8…QD3\D]51k%b5@R1=>K9;̕Ur6)fuk5/'#B9W!jW!"I.ѬݰxJpYb"2CD6L/{yDVX= 0/]0]} Ž=nȿ918zDO~'Ul^pVqL:i&|kOzi.cXOu h#z^}MjwNg:(gs!DUoR~R;Ot~QjP sMjw)E{v݁vRRS!ڈ2KMMj >"I8ed)S6T0~^Njw"2wBģ8yHDVQ\҃Q;Pf]v2+j #;kv]}v"A甐 0T1Tޝ^-<9a;k*-ųR~`'-G:Fg{^ u+'Z:Jwp|y"0/%-jqY5Cʝhn\gyUj8mj8-JcɎTsk.U(˽ S)%R>69xudWwTy_jnYqi)^mUW}Jhle'[P= c tTC]&n?:^.|5h==_/=#VS ҵ5)$hbI[q=$gE~gh!TH ҃\AzN`"A"B_z Dl_0瑺DaJt!P խ m,~-ȵR)91334TGلJA,Ut @)ZP`,,YR &tTg&Gg,EqWiN4s.X98"DC}hݏ ѥ-ZE*ch"'tbʼ<:mXEw{=y^10L hLU )(D9o@A\J-ѐXM@*Cђݧ{[GƅzyVϭ7N7 aInc[!.%q!4n"(sY*v4-zUGMRe2 GÖWר*jl] f{Qm bZ¢ խ(GJrp%&NerVT+€%HKT{hrt[7JTT^1tEҫՖeeVBj^58KEg['=0jITrZ#z~MŌ}d+ y24:UcTߠNU{P`*O+fyY]ʽ HƤ0ul*d x%VI >)2hD4aȡY݇$ Od^t9;TPGGXyiߢY\_2ti*s~eKѼ4x}voKd{S?8)/R؍N< e(yspNYč9P N#޵q+ٿ"~bGHȧbhiɖdG=Kvhʼn"Qd5.㔢VQE9X[+XjӗM\{Q˾+ kcڨlC0\yi ]l^19I-}EsϮ@48MLh\rWqV瓒Ĩ .J &ȯS>h?"ygJV·7E-S,a=1%Qj%|+?r6g˫+X*{-;a/)ֽBoDê&C;"-]{݉yIZnv>',@-٦Bu^鳁ʖ >p"ԭh 1=+IsuEuP>,^WNc=[ )gHhfm cӁ6:)Í?;X!ѳ/EewPdt;)o-iYk'1v4MN+k^*.;Fޡ[TT?DZ)hHBΪ{G#$WV &LHx7mrSU:*TlCN[Y{hCC1s#^O_XRۙx )B.=TdžJN<O>*f`,"ro5P`W%l>޽yp̅3SJ"xGi"$ӀoOfS QKyR[Չs$~z^w8X_'}&ZAoWeDIľX㞸}sA篎ONDj<\2J;$Dy5!Lb~ 'R@UTK &kf)[|xU{&2- GͤjP. |9 q ODk;4J1v|RtS0>YUi+)$MRj{hCC1wퟏw5ۋ_RE6&XN{^M*ۨi?jޤ|%S:1uA_"˹8u Rć7)V*gWjL øU6d'XHZw7{'=YJmVj?^k~| 7R-JZͷJmlߵ(dkK??Bxl9:/δҬ =;+z$HfGe=8rUY_= jE_9zʚAKrS=1_|9,P_~<Tz ”rƨ RdR[s>I"ra'-kbjI%@KL]#hAx9W$;(g$( zLU:QaJ*@,樖c:\l)%ʄqt!f#€IFZLWu: n+Ń`lh`J=1Xr?aDrL-)/F%瘭 lK 2م] Ÿ g u1цbCTZM2jr\2H꤉2ripGL䖫T _$H9X PhCC1ABWuL;¬b:i >ӫwr8Zs Ѝoi΀L>ؽ13Ks>&v[(:O'Ŝ'.)<ʬ?{yȢo ] Y_nw Ǧ0({W[.`̚ƾhtɺy;d٬.&P%[O#6vR"_&5m[]uQDE]EuQqTIdȈlpPEw g9qY](;6bgRd^WP`^|jl.l`sKӦƲyBUgU5589M9L)U;b&׏>Y|6Qur~ry\??";pp0e\<(ن_|~+(&t X6LJ@.O$19(^cR>K)x^"uO|iԤ#(gm" |̿_ 2]$ ,F'zP!Y$uB2-v`3R5cU,(JTߓD^|#hszǨk+33j!*Uu@Hu0] UaW[mݳ֖"i!*mCFL%F]OtAn -BGrCF$͓5-{+S*L2˔Lˎ {YٸYvC/ Jno ט= xtr]M ,LaPeAk cՒ\FJ~ yӣynE][幗XvTz)u:#~;yuğmK:M/m=a@ԚCrV|L"<,u'rvNf2"J*"ʞM&@II ٶo|SwPM8aED/'EXwڈuu3[Mwiܬ`N~NY6S7wN qo1/m+? m>ކOwlvmU|wk{2XT:/@ /.n:߯OOr~xW|Ir5xyJw?qw>M {y:c*=o =Y\xxJӡHA|/Es%xT1d2L6q.W%x7 EcCH9ZbS KLMϠE蔞

]T֢d+{j}OZWoIzKM tlwrʢSż"0XxĢEO L| }~GZkj[U~QM ;+=YኲO (|ETPT&K%7f`a\vj=m&ݗ;Þڃkuuv_|F$~{h>vR<%L_9d cUDkj1O~(CPӞGѱs:l$˵T[;7C*O?]FlgD?'>1?۷hc4ܐeJ l$YpV]7h~Qc1ʤ\V$nIP*iܤj hH?g=rH,vhM$6DbMĶj>C?c2;7">.#,1͐ dZw75}<(1Fя1~Q̏1.4xol.<9EB'mwN`/ilquM85[(Twmm$IW CZyK6żyԖl$nYIrKbmQrUdĉYQqҫEM9AJ[|荀NFh^_}9E8ٗX `q7ۄ\N %BS[J6k ˾7iT{S8d $Ps*c 9CQ,w{|$wb7}dCIѻ.kz^]Շ Z'B\ M}??Own^;>OW Rb7V߉guEZw׿Z73YS^w_^_|pOVstr}}q怒|o,LƳ9246~^w_8϶?ԗҗv2WL,1.>-`c>Ogwvw-k&.(sM۽ p^Jrч9UϛyNR~>ͳo?$+ }$?Nv3{NCL'3F؟ m`ȉs \3gmz:L\!kUb L[{՝'Naz NN9`-~ !"xKv|;"Wy_T4?{CNꀸS0f9->9,w_~Yo1`!X}¯ȧiBAV~ 'k\u^;ox;SM bQ׻3=3)kIhxfYr q Ex|zO*6Ck(Ge+YFtHA蕁܌; $x)-P-YTU9@FJ hDg w&bMnbfd/Hl,KIRB, w3G]SÏ÷$ '@'0aPrE1M@ BFB=jK5Z(,|'aD/Ru5cjbj;eq *_t1i+( ,bOOl 3FJRWdR K׼LӴ5Gr kREN+ Aފ־-Ilޚ7+-U$@ bY,s-uYb-(!WAgENrNe%czGhc,Dۓ@y-2 ՘u؉X-my?P Ha.1^*FBI֣ZJ,aYbrFteM TY&LUQ)ˌbj>TAy7q{)BBb LQQ Lu-Vk BxWh'iY I-_E (g \JSJ9Gѩxqz#(u|/E$Q 4bI~Vl̝ E3ԽCQYZ/U3LiW&m/c.T[ooYO/wQXQ] n yy ns91џ>abE?ټ5b?I%2IDhInpcȉs#HThe{[܍hez,ubpc}^⣕i2My7FV&GIr; -^  a2V5X8_Hm] OO%RVl[ZblE6R,H'Z-I~8+Flc-E10;BЦ*Lun'&>IQ($iQJH+N[Q*0B&%LtMb>,6`sf]ǨR4d Ɗ`Ɲ,4_1{jE<_q,Y.Q'Urb(۲Uↂ ,MluɩdQ?h(jɧ\XhѷH[l. za,DїuYT' W:ۍhkV M25]AD`2Ud:W*XXM u?~9;K}K'`YǍ5kkI&JFKt.Y4P%˺wvCMXXⴆ҂LR_EK-,>DPJ[}CyI#,aEU-[ Bom2MFʋ07)AC .}pA }hC'0P|.w~|JǢr <]!fSE(X%T6еZH ,ViAd(ŮTJIZQ;!ޫLN\N͖cB$G$%\(IN;ADbeL?D3*IEa5Zr(ޱT@K\v1h=LsRS,IOj 1mFihF]:5C$*^eLfWF_*@,sZzje9k#HBQnsh-kmgLKjkuʝqh-;V̅q9wiƠknU!**ӵsb}xoq:۽^Kr+j&Ȏ9݀xVJyq-b<:4A;fY+mfs>,Yj`w r}NWtC0!y+y%i? zR!HAwP[O,9;GwW~`[S.W%n ^wu#x39Q7a]. cJR/@vbDk`;DݺܴMfl+w R8Hzb gh]n 6O z w:=9~O0e=r/ i#;SѦvri>OB]E?K",fZɠIA#YBC)hf0 xqwfH%h,ZN[4d,e |Y;}LEn6U})-uW\H*vXZ0n|CD^E˨,' 艍݅y5̕cBv/rH"HvNZ tvm70-SPU aZTRva^MJ6j }fЍɞ6YǡQeY:FZϺ,1Rū0tmõ~|ʅ*t{kH͹J+Nڜ%0A1SH惕\.Y>RzdR:f'脮dKa6p[6C$ܨ=Ř`- ^szT)GARkwG=ɑ@tTX@UJTϵ*i˜@ !}G SēO{`:ntɱWk,, ͉`i'G{Y[p<ʜ Q*QbrX Qh'\DT 5\^/| Rt*ZPzUMC <$)Y -\ L1eR^Q$UI7& MqIجg2bEBh!"5BX*RZ'"j[E1kΥ)Pe>%Kq#Bm" MA̞,bs2YԘzB9HK|nӴZ%hQ처a[9N3m5NH@1:#c22`֘t!e"QtbPDIKTۣ6gbyiUح|U?a$USHZXX6X[hUH& 02YтvE 3!q OZ:B.M*RpY ԚMCތJ#ׁܯ.=AN١Sׄ_c] *7Z:8V!*LlqV\߮n +Gw4Y*QiǾsX}.2R>!WԱP10o#ٝ{+}t :xU} c}-=u5G%c$LxbhŢm x9_ݐɓCTTzT5.^:_. 7s.2_;+q2&s(2l5׎nX BO\+P{ tĚ+44-P4 ٠6%_ޟVB#aۢ|ʷw}4#8ɩ12z~֨YZAA |1s!dڀrVʥSL;g-o%7}|yV%4_[8' D.7@jCcx? F %q*6zyjݳ8 y;pw\@v=jYT} &hgrsG!UeCL#kșLs:N 0U!ZEfCގrYt㒃\sAv V|r9`!RaSK9b鞅@0NGHtngR-JsK`z.ӴH,[zzn5MnN д0ȓGsP]95QFKR]As֐AI 'ջ=M9i_2|#Z'^Z?5ߟNCܣ\ؖ\91#q@+w{ڝB@$ve4%R:%r>X 3.uͣ\ߝ,thvsAv̩#/ (SmB/_H%-U^.n f*72B/[eϖ0fg#矌QLh@Ww}[Wx\o~j \5#3;$' ˀ=&VKzP/Fbfg_ooأ/|L4fhe"8i :sZ̃ ;,U1{W`d~қdzјy4[x61=19; 甜tgPYQ֏Nd_|ի޷!$k>ͨJRl*ɖ Dr^!Z}^bVUc[rV:PWil_I!yyQ;)\9IRZAa`%vR؅o@>nLߍnVd\wF?BeǦ/O>M=uFWяNYZ*V89 zbUw//6PKxJКYYc,$cYl}y{_"k%dNd.滑"U cRMjF$Gȋa7~Yw|y'Ně+&4F!A@ԁXG'N%ϣpYH@TȩGOi8p}rN J dל5bNߒaH)/BKºx&%2BWb 3/dJ3Sg;ՠ*^sj(:Հ`zd4zizڹc5@H5JplԮ?rHժ TlaN7QF dTک]Ji^dy16#TR/)7fs--iqLXmW kĮތUؽrbW[/x| km1PF5fטfg6fB-1eJȇmԮQN|\NhD`-ac٨6 ? n}2?/N,\q?Wƀ.N?[O aUZe{e)WGC(ȵRAP& *E U J%bpDKS>hD.˳O7>TeK< 2`2HYiB <@PbHwCJaH4d8I<'9iS) ھê?G_d)4A\`s)&MWRIHsf۟'}+Mfl-8r_{ ӎ|t4}ExS ' '=1#*.F}0it$fR'J% yJX/h?hHb`^ n||tq6vZy<)2&%<+B-J͉!FyJPz2nETQGkmtC?D]=U-Qރ %e.2Xi jn.oW5ѪآYd^*59tCf8^O͗d!Y'A8E\I&$b+cFMT? ?Lluor&)C祧??oVqNukxE46E鯙kGBxK1BeDBD% ;> _2i[ ЅY`J9-L#-h8-}(Dmt%bdDdZ x"Ş?=3lPai\gKyrn Q ch`V9`f!vBt {P}cseb˜NU2yFDIVaiU<ov ݲr[}7RpN2:'.AvZ;hVMgz8v_mϞ׍EҀ;p Nau,[sLb,]fzXdjUU?Xsƞ53owajOŢEƣ_x!%II.F=P+T 6g-VXj͸C91;E*AhwG>z$ (z߉o8ȏC@ t/FA|=_}Ħu!SrgB.#6r= ,2 |9EqA`nֹl ['qn|nqI~2\)Ü>b+;!s9N20lA͢IZm5;hk8O<([ T{ׯM^TN^1fïWe[Odc׻'c]hX2*z6XҢas#8Zv)8JB讥UߜN#D1SO]m=Rh1X̨)BDjհ=[('vhR;vxRvFl]>jӌm@¼T)[ } Ujw Nԁ({ sBL;SNRn|CҜcv `- i`ļ8j]X¹V=vvaEKvOaD*f>fg F0 N8[bpJحhO*f''Hva:kk%vKzʹ;b'Q ɠB0tE`YvJ 6^?vj'jpR!WcHGޱʹSj\*Կג+S:F}*c<ےUq2dxg4{]mdUU\6U3>QlzX򤪞ʓ|.@e{A]&Ò&ަ&pp}s?Qoze$_̾=W^_ee~^m5|zRyR^[+o<өE\d? &`P{~PBp y}9 }x=_M_,O!4e.b04u mV$嬀Ɓ+ iFXi*}U8H)kULpOG}Su]k* \ud˶WrAީ\6^SfT(Cvw{3hfD-BdYR:+\5*n\Yu,wފE\7b966Mɍ>{kQ_6gڭCV(zVą_,8̶]MlC;yh.k݆.[4~(OYhS|P[[ ch\y…?a)ksDj ͫ#Tǿ~a5T3/bn$,US@ˌ|, BgbosoLgO\W[/ѧa{Rv ԕڒlI4c@mHތ?ǘ7[f̾1]Pp@#jDI#J&V Qg`Ӗ{a-v-V-[-6]$PƱgCC D>؁BH. ȇF>@R:մMKKbgMK;д4юmu~?M/kOXT %%Hօ$4B "@!Abɍ^֐`1Z > j2yB$o]hϙuhօ΍\]c3%wh <:q;7 _<%J:Yd+(hSiJ4ʌ= c, ơQJ@`T Vw8ucd$vJ+;rj0˜Cy;1/*Sj<RvR0G"nPŝ{T#oCvFd7%dM? W3g˺%/$ ,/fy#B]U%r #B6+ M%7&@\UYYTpy#?++Z:I=*w5Wge͆ E"2RPBc}a`)+1TEn&8Gnמ!V,m~N!@X}bG>Þ_F^WG"s7VSڶC@o;gLԑԠXs!9Ec,`tgWgl!jɿc'@ 4̓S:$< JF߅lRw{BIݴ7gwL'W Ϻ֍:<ܭC&k4iI7\m>?$5P8֚?V,(V݉o܊Ri))X_o<u~/^%V jn%!؂> #j 4c4? 6ZݟSuVUJ,FvɖYeK$.AǂuĊ6V$Kȍ0,e_feukoc]|I?a!(ccឩմn@ Hu.vgsDb߬$ Ђ<϶ {dЂAC90ˑb.^XɈl&eŮ69䭛kڕ7v@UW&UaUUA;kȎmhjENjzNd\*l Z_B  >蠨ɞRtP0c..cpO_ jɗC fBeK5͟528Oɳa-uka,=IKW *}{,x0Y`]{֜6#&;V"7R*2ݯN隽7X7ٗA63fuJH^k!Ξ@ـ z]]ûQcAp]N\ܫ-ӣ0jȣfaD*<L4D^Rn|95rDҭGvJC_.yE-?/G{'Ch_qcQ ǂKY؉)8++wu޶HWGCQeU!VMbIuCaSEP MUTgKl7f5|/07uԭ۪A]%M7b & t\I]*-*m!X:~z due|ZSF&/D@y(MaX[:Zιyqc뺞NÇ`Um*7Ob]qtoj|{$_oh?M??k=0e6!|kR v~y#̙Vn1oSzO;|xZ;s\/ klI3nrך؃Eg,q֤i6O ,!G ԣ;vX[g5/7&VBO* e>5¼>Zͱա_}bagmA\_<>O]˳-$/9cJOEɯ ;XVwMa|ۥ޻?Z\-'߉57=R<'hm-nI(GM#=i> pӁNV{#܈TLGD dC4zSE= =v)rrt<桍Hl6Ti 'ci_Ӫ}v6~ts ζûw2RƞGc5s}w3H_>@-[ ,=ӽla~U_gR~&@J@/id̉|Il3bV&vVZ7֑A#W2$(;m0o1{K-E e*٫_]Ȫ˫S`a61&\08Af2@á7D41CH ︬LX74:Љoo׷ R?ݟjrbgy9"+W} D0ɞ+rӪTJl|hLnb2;}SlFԕзTT_V7K7\DyXuy>AbJbgZД.y.5Ya}]'mѧʶK}Mox/(_Wjw`\փd/{!X'][sF+,dwO ϭ{fT-ۛMUkl6)$ƺ^ĩRHH"KL3=_E_0' 4mV$<F)쒏$e1.I&F4{ɤ=D4+swjV/k:fA>`!0s/<#n0@-qOsim͚.׭1EݚS Sv v 67_㺉(|cH/[[!+ ~(.Sx}y|8Й;RZΠzYb3܋:'! H4έc:+n\/3ѐ5̭q3 5F&qաvB!DKԹ{%<ݬ=:/@%d珊yğLrj58Kn.4q$bU8.Bh'ߥĭ5pAõ4su U8<]}8H-r.Oo } -_I-4c^/E6;ޮY=Q椤-w0qɺŬ/S;E5Q06u 0tǑRPJ]?6ŒQdpl (`'^Ӊ[(FAHAӥLʅ)\-qef $<.RNKҦH͔<랃|S;4TL!8[ D0ߺOb#:@yYnj8(b;B!VXTv8nͳ idbh\ť^#d%j9[YeL\"KsS de~ z͚; gB>ODwy` az׿}ۿ'ï 3]IoIҏM?n,?v%޹56E?<ڴ.@!1ס' C&/ *rn0=MOKp:Q[B%teG0lL0!Nc:{9V .F]3F%Xqvӌ|1#tN[ .\exGzsUQ.,~GDȋa2%<܎#P$6K+/xrve1ѣW%pM]>x9yn~*9f=d?b&σѥ(yposID7죗$xD&ueG7Gшv/(~1prov|G18{Ft~<8ge1 U?R|Zc 2h-L2_W|Nc8/p{?vԗݏK$g6;.Ǒ̵ gf'ݿ_4/YЕb$btKemLm}} |ȫr !r< q9S|xKA MT( 'i fˊj\<Ж-}eYZ}Z4WkTyB ʆzmA8/"?̵!ʏ2-m60U ,!wd~R.|@)#/]/{4CbWYjmo>(t\-Ui~w1=K7 J}j:**i:|m8,+ÝrͦITq]*M)Yi )g Ru*ERh3-qz7I"$./?%5h-]sVO/'XsN2$`TS'F,ˊKPBe8جt&T0:k2\kTP_ﯞ>c&U)PyPbZ$ߒ_N߭}.  ]`R}0V_}t$_csQѝ Z::| FoOxALMZQTBj%<.M! [ۨ7!4䷂QK)AiSQ=*:L*~3I 4Q8yY_/@!J 2Co_ FP2- ТC iqUm59s(CM(d:;s c #B 3@ nKU/A> C9PgCp Wgu-}EsM)!lhxٗbU)\[X`΁9o|TqzLtxT5?Lsq'M \x,cL",!9Ku\@fgb mlLRv| 1j@b\(UX|WX|\a֢nXX /E-J\ָH&,I}uԒ"q2ϬKgJdJTlVr0,6€5cg {\k ư8ujEپs6Hnsm&2}XM>rBX+'$[q;Ozzת޵T{9@+R]v\GCƣSKq*pyu=bŵYS^6Ә 6mW_|( ]!@W{H?f4~8i0bB߱Zۥiw4;<СnCci}u ӰF 1G.",u)BYKfQid@J8 '8487T(ByE(RA0'̷ҿO}#n~gI?]v8-r, $HIDѿfrnt'є^]Gua[~}5?DU tteEٚXܘRnt[p%w8IEH`k!C\d[7%Bvwҋ[AC4Q=t`%SY=gV;}KAaJd(:NMDV_gӻ gx=Sy)p?HP:--To'VNh+a!0: 9%IJUɵDmh[l5MWa,hc2VLQ"}98KpbbUīKԹ.܁8]jyx|Z?X3H6'mh)}܈H#fJZ܈ZJmuZz]qN0[blF%I*J*Q"c;6}Ld)u܊`mmAꌱ7gzC0B:4iOӁ*vBfެ>C 4%C[<5‰ ox_W2KuPK1K0+Oz_|K+E+A\nik%!6[}J޹;W>y'\wf7H,6RTNY*3gk1+b;:͇KPACzMT4WOF=*wъ0>z+&W n`2xR%x𜜹`Ԡ91OXëtJq*էԮ%Ԏ&/9KR!V9߷ ~s&2R5 ^79gc-Y|z׿}ۿUGu;7ON~6 Iq⬕Ӽ*η,8I3lH0V}f-lDY+w.OI]N ͻje(U/N/Z5]@ղL.QF| tv{NPBccQM'ILKn5H lvl܍ƓlK}!y ('e{J環/L" } vh?Gg a<1uWqnT\q.b]/6]۠(fTx\%u.P&SNJ ˬGDiB.nMm6+}VYWQ{P>Jy~ȆH:Eb(+'N4-d$f=5Jv[S^at^ k[K 4ەtWK` 1׆.1SGT1_R"롖d;P>ohGP-ANݐی1:)5?K"IQLW 5㛦OpI -ɽcN gԂ$f5AɮmY ]pSTޟ/M2p0#`":Jܴ44;7E5JM*N7AjR%j1Z-0Nw=Ra*@|=来A蔷~W?CTqwa(i`0FJw_o zg~I_U>oI,/ʜVpMǯxerX,ue]>]J9mv B OocF52{pjp7cR"1 q Ȗ"W~*+^ p{P&zR Մɂă%&'X];tTrY KQ5"=ׂДoߗQC 4)'^>dr>0 Nݥ) #Oו8J?:p ,3r]},ַb},ַjul-yc|/ZHΤ\&aueE.=`N 4_SzNaՎ6Zj6{?i'o^ݝ@w;blFUصY')T|#ch-%,5fe (X{0.u0(Ĭ㹧rhD9$5ҋ O%G,2mh,t;}zů JvO bFhbSD64 jkw4 g*]`Q Q Y 3FB"ZhэRA1^hyNРF[~<z5)>8=S4 jS$Ɍ<ɬBa\Ɲb8Cx(gâerd cB"8(*89aP8iBђ є͢ݓfÜ^;Xz%أxb}τW]u31[ߗ)DZapC39~U>f=6PڰIxJ*ABKW 6pi}{%εADiyй3T .F[m&iapb5yfTHE@ H-a@0gŽp^y<-/b @ tA J̋%ד!SY]IfZtr$DW94r^]4蓐'Q˥BVx.i &P"UcS; %TPH2/6kX0V{'*fĨ rs2s&~1qv3ng:Æ-.#:(@X!O'cm GZx9Rc =0:`[w,(x$"EzjGP&YX<|>ǺՄ뼔+J37b9ݓNYl3W &#K ps_Z۩RƓi7!Ҕ>~K!'{v2kEmU2iV<6mB$ꝡsӔᑡ7D)wblx6r:Z2V;5$mzDP*8c]2fGC"ma-,1{L $%(eZkHٯ& [ ZcH\KT$_kJK0d'HBڡ/a)R WCi>+8̅ϱySW/%].SNlxtdb&8BÕaٮ@s}ϜΪuwv{xoTWT#`zY]] ,D|$BdrG$q釮vM4#v~7Ol)7j[qUw,9MMHc;PY3>e,}?뻖m/"RgS:+lak [ߊ(ӫ'~e+˪YK87J5є>הH70<|4 bxx4ķ[UY4!dMUUns=Q~q2oo`ΠTŗ,(%j׌o@v{N8Hg1(&D#)\'j#'`4j,-"D4, F+҂lv+͈D "Aiވy7aT+d{ R5wKV%$+tZ"Мc&//TjxPmʗtEߛs͏D ZC0bU\4\?`Lʍq#R?h3#E6EņD=34!/fx{1pA;jHD#f{),Ua":(!@-h 9ɕw.@^0^ ä@E4Hڣ+T4E;?E*2S>ZhGv4Scϣ7޵lEds^<_r?hA~\Xg߃S=x~ 5:*ikB%[کPZv5&*YCvR.S]5.#FliAŒG<&|31A|&)%`jÛ?+Td\;w܏s@).aŪ{KuI1V@5^ T0 1Fs& 0iCEPv#7E G9T8ޜ?YϷ!F[9 ma:-eRʴ0,I!j tU~|fC>MH< ּgF3\͙lwoko˻x1Э'AsDȬ\h-'xǟ/wʿn([]ʿ1u5c@ضu Q[z߽ 6?t'VM$r9_.(}y[tOƞfCof tE|MA`GBRrS(CWSo97jn8,9yאKF-,hkr~σ+m#Gvg2C?%`^$I٤Dרeg|aLX nLT>@"De%ze@R{QmEv(RBUDIeMɮZqP-3d9$[͡FrP=/!jNu)0!eĺκ٩IN%*l&g8ɐ0+RUыsۙq0|ҤԺOl]8?]m t"p!I9@b;Z:}o:no~7"Nפ B\[cƯkZ<'FRDUrCk8"9aeTZn uˮODƔN=t;E[)frbv գ]( $ lo1 % j\SƼzc qKj dCDĨ`=nmnEtw+riTD+ܧ@;hjwku@399Iͅ,lLDG%g1Pɏ"s 5󴬨#m(* cd&kc"@}ñ6 1k q9*>|h3T< ʯ<"Q%(_YeԶ"wC32 "4KE$tؗaѻHc+qLOEN䴌ZfHNf^))X2+GL;.Uڔ @V)b(;,{7ye8d7)xAȳ^-"%[ٲhOȕ_,Ы[8٦s1Fo[qcO0 Zbd5sLuBvRFk*. 0Řԑہ!][q7 kFr8"dDrlkj qFQkw';ǂ7v+x((7eCbr]kࢫ4o~2{g-̿9X=4|r"+dNG%\vc4+d/cen}j4]{rAB$HDOIboQK"Aqa:-0];5%í/:\&72uqݴC0l:TƝKN$U&H2.jKYQ[J*:-o]ewܙ \!T|]uEml=^qG-Bxx  ҈R}yxЀR/ʎdQkp=tTMYuH&`|Q Dʋ_'oW!7L'6|P\9t>.T+d.ݎ'ׅswܢY ΠzO~KO<(Cql89[8T /,f1NX$gE{! Xݟ'eBLlvv΂g׫l@?{WU+6pI,!2οރd> D3ݪS.. tjlloߑ>ЉgAg.? L^"O!UY;ټs.[\=[<vv}y}pӏ'ix:]%\ m+=qٟZ??սg@%А]\/./2.rxss%OChpfpZoŅWiqi~w:0F?'@y- \ك9]³8J Ir1gz5^TP0gۺ* o_N)"]\]W-ŷn:쏊_lu/ 4-&o/ƃ=|2Kifbҹ~~( g(^LWq/\VЏQP3˰ zyǣs¥_m< Ɵ?4fPR9_Nǁ~`M?Ɵj`Р&\0OL&zpd.`_'P )D0pd/f0_/BMϥژMS^ϭ7'[2KA>)$CY_=xqWr*oXl?즳yp#88{uvE"888xCI^]Lv`0Ը[%,fv,3;L|Ua}X.jޅ%w K֑uސz!gL(\/SZXШTVz<]q grJ-6\>,|at k5\y2Hz/wJ @YXc' Iż'=,Pr.)~]rV3d 2p0=Zr]n8_7fN碁9S"‰YL M1j_WN㗶`+r{vzKa=ѱW|7A#@ 9ɪ(s*I e )  @Y"|!X,((o&,{$+irpuѫG7.•7#&2nxzs\sq, 51vz^H];-٧8F̪ұx/ щV1k1V3O.i45V4i3DpYK2 GDGnNIa #d=R#H2"=|$#H |ѮUMRJdŒ%d"3JfC @H,dg$Kz5*7sAGdМV;Ϟ&C1[lW1V_Ud"0xz)ʽ%0p'S 8"1I1XA*RIS4'XjE|iACŐJC cWܾ,jCV=+7+SXxuK+V_wٛVuTZjgA;QGx:\y@)c!C!éD""eiJ%Q]?: 8*$C7md  4YRc1N%*s"*$CHgU1 DtpoK ] (_kxCP>?ofIȭm2PY G-y MQ:;#z"(%\[ BZQ(Y<41TJF% D7TG*mFTB`&UW@Q;`r9Æk#`s ׆ ďBҀ*`6@}iDUnUMhT_S͙0{Ԣp `Rc-=G6S|;i,W~"~'kو[לK[`%޽ Jѽ?]uۆ/Y 5nbN +ǫlOOsA0 Dgu N?{drL1{750ӆhO sl{cޮCydؖad\Wpw((5t(|귟0˸-ۼ@᫋o>\}o~{y?zgyVgQ[Ex]s{`R07&wbkڰĻqcj!eJ|#jLL,d@X)~ĕ2Q_G|&8#|ƽv]08yμK۽贓#KgήNg …˯V#n, 2̖o".7֏M:>}׎>5(O>^ 1NE@hk=Q#E ȞcN=q 4ĆhOh'eQpђ9j")23'ߒ^j4+Hv^[ӳSr-Q Ҹ( 䁛 qkCU" ʇ*i1kľ6z+ 41ut DEGr?6> 98( T (O"CYWːBǀQeu*8 e2QtepƄOH42D6 ;9н$E`"Ό'LP=x:,Ց3& &^KMߑl\*/x<ȗl!K$&?J"BP: b@H(H TXz!(<܂7<aHKj< :a,0}fd (d!J$PpW O#+@ C V,LB)BQ?!=\t3A \'Is8!ۇk<+Ns.KŨZQU}1`Qy?*2!JtJD"tJ$hLKPHkc O# hFP5H3?(R2Y Чg*U(Mvm&hȳN'uJDH XvkIRDIX_%-RGIk+ ڡWj- nڇs$V?Unhk03FHoelM-S:D APE.Y``wevXUf# ,ؚe ehPI:TRN" Mڡa=+aРq֌ :-ILCDQ=, N'{M˞xTj Y]ͭP۹zJNL[lO$@㕠E~O @S =>@mf=|l;DcK9olx;0w˂g1BK=_8P&x6 //Vp|\^Tx>G(;SLli k2REh4[mZNr"x䶑mkZZ*P3n۫@U\qnUBh4yBk} q(ZzZ.Ze@t|[Yni*gc-7Uh%J=I6Q 'M#PVL/+ Ue <7ks:p]b|f/i^Z Q{hF0x041DB+U*VM vκi-O\ڣH@ 5I,DKh@+ЙƨH1]lx諈2aY]Y/~;U!.[vԺZWMٙ)³\-YʦE.}B?/yX?6d_ ERWjH 1ZQLcj HiIΙdyZjLWiD̞|[%%[r^^b[zpw囻O>.}|yNZ>{>\,ܧpD/o8,ƿM5j?cDmEG뉧]۴ovu.}Lr@R41$Z(d#ɍC;ܜHB~"Z#S䜜NҳnBvYкb#:Qb1nՄZ)$W.U2{},[-9S%m(Bж?u&4׺M!!r)k"vr|ђŦY'MϾG+ H:Y6}o F{0 9ՊksIN:Ie#> ăTŞ~˻w;yd@Fi~K2WopagoGa2=^/dO+ۿOiSr;SӒ 7ߒ>f]Xt'`ɨ[&&e5,>KhSdSDtƱdfX),͒]oS`Q!b-8~⥤+uI4^7i @M+CӣCzI/ɜQg o繪8ѹr }< ,s-匆jJtDž/cM&eIdiw ;bw ] a 0d8 JK˿0cdώQˢDg&R55Ko-J(35 ZQ Q9œ\Zy+WO'ъmqLٲ3-3FVjeG.p41JosNs DlcoӖ !x%sY3:gZxܸ_i HC8]{]#v%&ٱIWli=Fb1 Ij5YA(_tn+TgyCZ'~3V 9s= ?)K )P!pjhT5H, kϙM=giDv4cb}8srOlf=Uj-LJRG%Y|ΐҌxJv8C!rG?FxDŽ'p9C )ʬ)9 ARȂ} ! w05q#Y^6JŕW#a-$$1t]lcMF2wIPe).ADczJ LbƷ>~ԃ_O@`WBϾw| oBu4Slv{CwJ_(NQL('D\Z})=sF10ƀaL:}hY=%!*ӼAJb1t<#TB-FJ'<7nx~RGFpXdG9B&xoVw>H oe8H o=CHpnahq$@Z6rK)8LZj0ıh]-fGvfޣDQG4^jXJ; 2Z46;Rnl30a@a "6P{݂Z " T3YͲ1f+rF.9Rt.?Z}>o2D{3"N2^Jn/6,al(r^uRBYR*sX,#*RH__M<ۏpZ z4vO0F73s_wp$S;NbDv() t.?@@ZZ ^.y]+px ' D,F]M^sŁ=Uqxg j2Bό 9))t #`&NJ `e9=!T.HoSyfx38']9LU,nL%4 Oo)EMք_Tý~k_/ldt,N[dU[ؘ؆IP'ϒGż󹅿h6p>\]&̓`wN|~+bXM4\arR˝ jͷi᫥&a @ާJqD|M&s9H+(R3E4JJ0*-UD':턘nH"ƥ|ckhBj&$䉋hLquX]EhXT6hb)nфVnMH(b&lj7Y,i$*mԊLO$=2%rdڀvAщmv;ߢ`J>v&vkBB'pS%&ziN)1m?X&΂$ 6uj\~y@<=quXi cRY5B3ygbz()|N:@ӁE8~Y6u,QDzθ>c3ъAț Sh3D3p0HdLy\@">*?0V>^ 5mO 5}uME7~FR2 Q!a,/n@cbiͯL< _ iC}秦ib2fQ,$K p{p%?&a[3/Fͻٛ3?C?R]ǻޮxr_Or`lq2{wE\HcoOk-~ 69l;P;z+ }f>\^$|WбY2Frc ᢧ3*%` fY&a1+30Ei|ZI  ʹN2ꁘ<-l :E7k K\FŤ-*0-0J#2+}k˸8(meT2e3!KIBjp34=I<1өO Exk%:?)j[ː<"n],W-3dq:2sL"3E*Dh+ f६K Z^yIdF-c Ȗy:g\v鸳Nuറ#9&P,`uRXEa=cHXr;o.?p:2Mwz+/ݾbG.LGw}qzNfJ=DZ Bۙ/~ -4yiTDlvӔj,5<%%JJd=Ux)˴zoaR ̢bGvևXHqh "t7e%Wnga>//0K5"De\Kʼn>nvxKJGCZSRQмfFTs "<^am*m o奺8|\_^Nk`҇V὆x]6F;nV !1Z<AޥcZZy40&9FS<zjK9l:e+e’-prHwf^`/54aҋJ*5X p 6 ԂP")I$ϰ<iAZKa<*X'5+1pչ2;iaHHUR=AUa/BoٟZDŽC:blΛǓ3ѽ{/Ab}hcʅPRo iId4d B!xKqlɲW#cfTlhz⁻\mЯ`o[ *[3̋oBMcV%M_~}#ih;k=Z};ʶyUm^p6EJ/e޺}xNߖLZp[P~4].~;')`pXz/K|i%~Eu-ܻC,-Wo'H8fdǝ:0t…~HB \Jl`FSJ'qi3T}Yq/kàHʹD$e=ߪxi0,"}իW9~}r,!9hQ|)bCe͵490_io:IV:ITߋT؟m [ zP9Q%8b1Rl8Խ+(+N<^uۢ <_+'{C)]dK*8 *(%c숴ʼںSk+kZ:|rQIvMukYWlaL}ǕkQY֦8ɣ`EQc 6yū)z׮=)5[Y+ ^UP{TZL߮o] D;%OovߟݫJiW.a\nxwyݖB?qOݎ=Bl brt91 E' &6B[kcH邀xKtC9HHFqki.h`6|GC9\lנc]]s&qsZI-t+bպ8\%B&*B2,ߖ~A teKlOmY[ݩV3ymS}˛O@KEٯt3Yw=z]y *F*}.@h6W̘HDea)Y s|~Sb #l]FPeiZ%:"ʉu.^5dC{$2डm9{jI3>ww4ՠ xiIgى<%bb_% BB!(Aą )چa8\`(Kƴ+1S7ЎZ ,֭ {H~s 5F֩MX[ai-d*9H׋p}g<(u):1\#B?`S8vFW9LA[dDUYc&9`pt;,Sr;tl~Ns\dHd4玙qQ@+8638xj@Oi^xzW=[5i6̾hyij<gbp u/a ᱳn(j߃ߍ i%:'ߍ&4<)bf{J ݈9cX(|׀E#<>ht~Ԣݗډ86SQ;tyv:4\CwJۚRe NhOm) ;%Δ*M.%pA4t:%rzِ&T"x[:f䕦ZZW￾;n94$;9&nw͉o0 ,y(_>.SdRĦ zFaOVVM%%& .93a~'V,&,@*I3b^%=RoS()^%Aڞvz|QMZS5f jF$:H:**%YxG+%b`d%g;,2P(pžӮ!KӢ3o(Bxo"c,jFFb*ߐ:WRr)u[w4I}l>?KxGJBٯ9ɩ2@POBӇ yadHK`_9_u\HuoekN ` e8/IzlG#0_|n6S$֮[/Xi&GA\Z?>jfwAayH UȝHE܆Lxt4m:Q׶QOc;Qqf+ĉ{W.9'xXɃ,^"c}RG4?,>Xfe\0@7= x+CoE@ڃ..B0ҡ 15Wn0,lz,ߋrʜݥq\*B}2T*D.j[oCdTI۸wP ќ8@(sczݼE +z6MQjHAofA{U_琓H i] WcuRKZ9|"mJ0wvܠt!U- "E<`Y+?Wݺ-*d_]eſ&x 3΢ҟFP0 ݴɱ%V1O8ru馽 &A:@>\\w8ToIN)qӺ(~ʅRϩsWHд0El/U&h :Z)!H-Z`HlH'@}f.w +O1f]ͮ:3P;(>#9cjn_'qcFnݢ y~McUU,TY(JTGa$,(7AX/N{`c@N7XJ&08GDe-:?GA469qQ)%/X_`ʙ\K*5C3V;@ɅtW>mׄOXa?m @\:f$l`C}I /1b(x|c> >|5]%0ۖ1|/?u+j n[~Jg8\=.JF e>* pk # Vu,Mah܎xuk|j :- UJ+ɒ|.u ʗ_J"$%n^*rfUMjKZGj`uZ^}m$’3D*\Rb?b\3 :`pܕ*uKA@s=%;j즸Xj:e܅g 1(*z7'Q5^a:Br쨘| de"| $.1ExU&OؐKfmJ d^GI{0QIL!jg7Hpw m5h{pYPL s$3{?Pn= % :O)Vތ!A?'1U'o!4뀜2C:ΕbXKbU^Bg&o U^hZ1p3Kdz .̚#m1S L:m+ӒU6n_)mfYr筨Ե(FK8ĨB^c!͕;'y f4GlFYQX:qeZOʫS㠖+W1Y=1Ge'fC1(`\f6W0%\3 =&l+j}eRak-Ih$LC\}&Ygs< ='|:\" GtCǚ .aOYxwDi}K䟑9xfO$MG6=v@,c}PrӋah摿c,€&'b)+Oa5TO-ooF鄁 jLPT$jO>A' %KQ90U#J?搰7Y*jnC 'zI\ mNtp5G6BE?: ػ@uI 2KvPJC00 :Jm f]DRXstTD/\\j Ů:hʛe*Cծj+AP8iI/z2a K̿wA3kS*L֏$xXd~YYgFCo${>%6ޠ77w!T5^F Jv8+7WL~dgyao׋7[虿^o~^#NSߤNu6F)܂r|W$Ta>ұBIIy{ Xe^x @M89OPfBX}ԃYӤ\1J53%UoEZ#{W嵿F1=]Q47zL[^GBz?3/l6 Iυ xfK6Ꮿ5hu^OY}y^bqjӯ|ܥt8p  a > >Y> ÿ506v9-Û}ޞz+0W禄JA+_dĕjV֜r'F9L, 87z H^U/w'MG+@Nѻ*K!WUP%±<|ՓeFq{~Jl~z,(x;`1Hg*Gy>OcdJc[zOŜ< f^BGJbNm^!3_Me XDP,<H."WK!Q滓cN }|5rH/4?xv3Y1Gzm:|9bss~!14_!esKV*N9 mv{/a ϭ2 002zUYر]u [wQjU^?fp>*=wtY}*sa[8]I}q{޶eoũ}? ES= ҤzaeD\I]REIK-VR'A`S9;E"yb/FG@6KrQsڱF^ݲ`mJ4y-崦k~R/8dLv_B-qS/ą|d®TK5Qal⤷7(l{j5>u˱N8$%*ׄm~2^Re t"$*^ٲ{3™jriӎ'0Z&xS2 ˁHFj.1p+ք)Cne4F! 5LM߿us_&`<Y˲5:K[ua}-' XI7/ѨY C Q8!_>f$A;}7HNQKțO6H'c}o2-~ƞ6rgSt/矓hbZ|ŻGަ5pѴ/^G_(qB k/ KL%۶+&O$skH fo4Kfݏ̕ꥰ 5$dVNh 8mmb0Q5cC>=+-r5\@uWord?t}1#h9)Oa~ِ|~?8?h` Q~~ QQon W?raU d & KG?] }S k,>}tol5%lo֛J I+h4r f  4ex{H玧. }gF9ǖLwϥ@1fK,w=K\ZUFqq; д\*ҫ%{W+~/Ha$}+~6&\_4 xnS?M.vWΆZl+? cbpSSol4)`5 5IB)ͤNIfT̰2ZwA%* 041[ T(4UN@P()@4CPh_BP 0-MmcWJΞv5 IQ;%$ !?K7d$pIGƄ'{ߜ1E7 NԧrkmgS].5X08C3.?Ue`y0ְ ~G7OsJK1Kv';vC̥]V=F{:bV;ғ; ؎oO)g $8>Y8Y`DqN'ǃuF= Pmh4ADDkԽN6͉I?wABl+TPYՔ0R٘PLR ғ-VyڎS2 jSadLg8WHgg<"\dmR =ѣF8&g3g 蘀O;f֧5QUBkN s'rOTIk dͯoG?zP*x`_KAdļ2 ܂R5/&{=c{ Bh& ńFLY%Z!U1xha(:tfx>O9?ڔړ|ƨzmԍfwf2Id, Q"E0g:$A,2:IH!9 XjP6"~O43 J*hKOψ;+u8]zh=̧Ղeu%_9GYݸLWe˗Vr ˇ^Z@ ru5Zw(,^ޙ@ywGݫQ\|`<})z\]qԂߙ,hvֱ['@lo.5HRHG t<;`y<5hwm1>qID3 t1۷:wOE2+ۇ: `9@1hydW$_=1g_FO2=Tbx$}+YrZj%dc9C1Wkg5t?>QCq^ڟEx<&܁g G t<6qĦwD¡DHV 隵M ho#O|B 9:<7'?Rczx!p#lZ=?GXJ#rG_ bٙM QؕȷOl"9c1r#VZHgEώcJhE'IYW bλ5%2R]t:ᱶJ0=mh,FC[faD?̸Rǣc|+"6憯lᘈ.-Xxm 4;>q$L?VQ"(c 7=K귃C<$]䜡A I;m[o 1JNFw5In_Ă.(3:!qw5N߸3xȽPy~V-N.-?A> o'Ǜ:ysò-Iq/vРk[}2(/8iNە'h8<@<ǒ@k)R|C'"PKk D 6_p'\[P#0L"yr,$ K,9ƹ3 Y3Qu@GP $ތD  #Sr~h OaC܈jZǓ?HtecGF8~Zis8v')gln;Y> I$헼2c?U u0!^h7vb1 )roz Rz޼]{o'o7N LI 7(>L$[$f,ĘE(J}% [Ք7CD@%\OOCwm!bWl4<ADBkS0oj4/r̃$lיǻE\%9I)| r h+( x2l8 >o+ NAMXNn hY!΃C,oc_?Fcl[Tf.\=$ή>52g(v2xᴲ&q`ȍ<$\`P,ʚR y0!Bs)iE)'# ,0S4'FZ4$@B ȍޜT8x%iWr lp"VJŪK+@(L2р)TL0%c;% kLclp΄ F*C Jf A)V)&9 $  ȡl(jw凾l-mm-MHlz1cHl{[`0+]'{Otz|mG2Zfi7kn]vԘ}?q"`F -8GI!Er'yDWiUYW.+]M+#V)ECO .MsOx"o^d&98zQ:`gEeWlkY sz{UA=TєI©9&-n>UW? zQzi'`*r#Xt'7nyodzvFu*:bzOǝ3'c^n>_r C CTD"vвc}-g-wLhܡILo |/jxISx>r%R0O#oNwk*ؤ1 C`]T<>wVJY`CEG2v޸p;|T!>dbXk#\19\ 2qI̸C>=W̘A֡UcQ̪;n=g0C]xo;l\9>ll.JtmF,}i>GBQI҂Gđ9GA0wvꨥLW*:>9>uRH2R(RFPP%81c4!JCI9fZ*#DT4-Nn[HӲx+eue '^G2nH-5/ nXI=Kז-!APʸsCdtg<ݒDltg 9 hL#[%oC(䡮ddgNN~NI4p{S'!!p=^01I7j pF׿ԑaTSkF~ 1kP7:eH8R\D+e,T !*rf 80Is|Oܨ Ӏ? g 10oa$S"m#GEޠmC9nv$3_`lGHd2GZdEvK#k`d,7dS$U*Jow93^ 0iVu`#Ѳ$.R*ug<261)PQ//$"(aSY DdKD*D+xnXȴ2*L  w*3dhwCa@8"})^ 2}:Nl~`Z;uT4 tx#@dDժ4f YGNL sc"x4@5MpܓpI\VbI-!t1$K@tf0dC|5;.z9RGO)H˺$ !нx}.2$*RkpIHW8?n0*;v: ǣ1=o|bwW7_fo]xDPy1#yC<郜tdѸR__/E@tLQGYb3|XqiL(QӚbxeڋ#ս-03A@5G.<-@T >O y7X*TV0Py^r, "VX($ȑH*q`5Ē(VSdlck\,O3 |,V*NR\!,"p!i"O I`BE !h{Z6J>ۿ:]Ύj- an/] Li>D`y@S-tPRZ@ S"L^S 6QC'gW29u@ L %(CFh8n+(ҹPnkfq|E 5hbMrJou,|c;pцgXhKޓ;LFJۛ}\k|*4wWaK.[V RzăOlg 4lXd|5%CZqywCE"gpσ+I4'nj6ڏv2\:8ckـ+{I/'tspϽ=aݛT݆*X-~zWCwXQ|{x~Uk07n~WrM2z1ߍU'|0 pw¬p[Կ3Py 8}zrBB)o~_SBҫ7炍[ɑS브/7kq l*7^ ;j|0M Fl,TK6>'GA"  %OTU_N<BvR&"&8 /\+s =$p65 lXyk4JO~nb>NX>{w0%ﶎü[+%sWvjz/Ut_R{8\f'[wa+@{N[uz/6k3 yNwfӛ1@͜NVO `FwzxCN@N SHtKRסoXBB P/4]?/eN6*DSE\-ބ<9nDyI_ Bx͡zt+e,eʘFO_jz6Z.wȻwkx+sºoaE), ,LJv%hV~-4gGAp#h }9:,|k,]QG5b˴WO>l {jߘ~~e ߬zc7릿esa)@R(L - L`&ce"?//cr}V*鄄 R`RbJ- U@IN*0EZºid)Tҫ*:I 5d'Y^sN }ii]8rę&Dgo_B] jakfdkl/44dpZ IgF/Gh6YkVCeukbn ۩ƻ \_2n-)gjǘ̇|yֺs#8&I*6TW╰s@s% $cC`,Z.C.8Paܷ 7'!rmNr01LHQpnW 0q<VU;r+ eq7iG݉*q.h97?Am#SJ H D)'Us,k`B $3R)lor1j&0]-3xaצVB}c+n/(тA@ snH1noYu-d0@!F`ҘQK jSmZ ']*wɬ"bܛW\x|w5ih/w+_>t`f.Ⱦ?< ciގ]!ֿ ;LQt$q{Õ^ [r'\ 2;UfDlYrK$!\DdJ|u5WXNlMV#*ZWhBCWE4B%>UpQu8mq}4ݎw@XuHBux y"zLQ1Tj2^َwKeen%Nf})wvD@5G ]9o!T6[xݬ3S βd2pFMꈬGyDl` t)[_ւŠ@k(e?Uhl07~h5+ÑHIm:$wI\9 !NHN I=gb֣0(iJsy WB=LIC]ud0%[YS}J39!-'mN tTS6.wXQex:)4&J ;eR,w'L+ VXǐm,"q^)餻8#;ܖ$DJҤ`t;8O{y0^$]|/trJ0Aw] "êڀiV7'I)d6q}J<6_~xA ԓ1})z\lUglJ&.r%2ɬ Xrrү,_}<_7&;F&i3gqi崝 $ ]։ v0Gx PPmr9GCh 7-TTptQB0N^'r J"18@0(5B0.u 8q[HF1-O(jZ4]:[@ ሉSk@ e8 7aVSN8MYjHg;ّs8tZa %9&w$`ȇB-lM*qlaa\1)KYXtAAv0F @Qf2˴;Q)㲽zзOee.2|9X+8nbj_zL;͡AYh_cL߁V 6Phe`Oef }a5˥Am<5{n-j1\].R{6dB-k aJu>iz/0 i1V;Gq0hc ;T[[s( v<7 F hh?]R-X*AaA,3L3jKdtX_MK,A+e4Q9nhOvh^43 2(7T[ADnuksNbTiS]ʨ̡ZvTmf*JfWmf, $aqJ5!|[&Q"F$~SQIk' \U,ʱ0"W nH9H@1 I TD>תT5E5Icy16&I7=3WI)$䕋2> z=Ti{eyHnuw޺g3Nz^iן~zxGлWq[G @,0G4_~~sdb<)Ziduw¬p[Կ3Pǹ wOOU]8}dk˹="r#{stGF)5N.uvzZP -P9NY9!Lif댌P% Ӥ)[FC@(z2.g<=+@qf]33w-7k3[}o*hv8uuS$ ~/ϢFͺ8e{:*x8,ߝO9b#ΐD|*4Q3.1ѝ!1j,@.˶75ܝ B2cmv)h% VcZ ^H[oOstA:L.$K{/u@ÌZABA+-:GiZ)V>N/ň pBgb@00Vd9$#<'( )vrq(!jrqjN @5?y! BfrJU7M'&aђ~_292SI*)IL9KBjqW,v-}CR dm#7_uyaf( gH6IɰaO׻U@|P}e4mCxGhKxv-G 0{F2ɑ( ^q08Njz>iO:LjFf}T]-E6^7,DnSk*7H9Ps¹ t&8!XF %Iq/5sQsK h@xTF-0xxrhMwy*mҾW"r"LhW:1\LQ%8J'[5'ΠU½GN=GQ=`9`r4dS,nqƆVP<[yWkne=lB%ijufR-ٜ%f'l=fF04Kh~VCS'ﯗN -]^xp@r>́9~uf&(B +2cR}+m X#RbB9Қy#1zGXj8 bN%Ԅ'\bAIi%d -$ Ji\JhcFB9A|dKBB^$ 숚w_Jkj2w v#z y"Z$SZNn)hRNmۈQT ݊n]H eJ)Ax٘a-#oj֫i`5,ljs_9nN@&]q{u{aZ\щRަ>աQj<씸p*앲a %v .'.σoHd~)}ӬB6<^qFW1wVJVJVJVjZcZMz$mFʌ s xGwaӯ7 DuM +ku7;Uex;Om_L~/ɣX6&AP5SN/W7~ǟ ˙Y,?]]֙ԣ!-3ؒ6`t]J20 ,Xvɒ<3.w/2;8kçr{U;ޅ|Q~ɳZK7H`}f20 ;X";S(9>%ZKՖKƓ'1I',!]*.80tRq5'BDp|xՍ%ct8',$t3=<0F{j,є.:ft#/-:pDtMi,@p:Z_)n*(lbB)R_ !nsCr^+aB9OX--0$<[Ü=kLQ2Q Js<9l6ׄ"=ϕf'DǮD3}Hi%'žHPsjrlrD`e#'I~}.%;S9Mg*L46T,Әc#'qS3̥H i"AckEv҅.~03)9~H5⹫7$Mcqz]iF[!dB ^!2YfA@`^FfP`\(;JI}8A$A6zڻ7w"vsN}$8V ȉ\zb kI˞mjے=quwbҸ3yr+ʛ_ussM¤ߣ'p{;9HFd5C3XyiA=jX|˵2v?8}g!S )6_lt& bR( $Y`,aU" *C1@PcgJ{I8 Yɓ]ъEkq8 ~? \bg UV#M$؛C^D!)rc^Kǯ]U录գ֕U2ƵKAߌDõL4bhG߮VIe_"ƻ MɆsEWQ1o{LZ}O]8Uf7!|Qt @f;;av?mrWnOԣFZ"2 ޟ^=! Z&kDRf<6A:1vUcsZ@.&*|'$bT"NB A Р8AR6mZU@&u.'L;&z|zHXN^]&FUϦQ.){ڝk%L{G*2=McBkQzmQqJjvTA_2.4 ] ]>OSa 7xY@T+p1B4@Smj4tFIsM57xx>fM([+.ZIwJV]Js"P6zcQeq2'P+zjOANá:7777;(A+]\3r`fY;mw69VW_Ol0\ps1iU"m>'Rc)|R>} yk鴌SOg{x8`R(JRC(Ad?$X S(vv>X͉hJ,VsRh+^HíƔ+C'vFJ&1njiLzf0nH'`k.+x&(N*}"M4 ~ fqwz$ dmQ`aIqT]s^Hq`H&CsO1Wrc"ƣD*PcXCD43Nxԭ_$:څRڌOE;!bA᜿oqAmTEJ\;LVneߥ t`vA_~5<緶A!NW]j7M0O_0ݝg{qͪ=å8ч["/k Gm+^tYnZI5yJ^FԻX`L,7^W /M[&"bь]_J*z-|]WPQYE1G\d}j1xХXsߟ, w;c%.Fx?&{|Ԍ0 3)ռ븳J2N%"ԛĀ֕T)Bv1I77( Yms3kI֐Q<&h9uAeT#E /MajqBK?5@('Ʌy˜Of6%|h&g3}(JE>Q)sM.N}!}OIm{gcùᅅ>xZ/`7WǨCp$²kndW Q / v\"WhҰIY<>(ʐ蜢T$-Z>Hn6]qaHj>y}}c,vݹf~X>}2><>痙3y);&Y&΀)n5D.c6C>6vɳI묉m&7r`3jx{s#*г7Ԯh}Vqtv0bۧetn2 -?׏M@USBi_%i)5'P%1|m"OU~DH*D<#ըn"͞%WiR_,WL8 `^sdppmJ& nqG:=P@?m:Ũ fMhF6j0Z01sc t|XS)DP;C-{91U ylկW|U+6v]&KncwdcƍΌW _ ,0ϳd1>fA~Lg_VGuv_:?.XJP@${YEE(p@"uF;'P?~O'l~Cq ǜ2 @ QOAX`=1QVa›j]?f5ԥ%8 A'k8\ypLjt0JF.xFLB8sq̱*T! TZ`0* e:^.eX [Y7|J. }bRtJaVb-wx,l8>+p)4 pmm#H}{+0%3Ё@f AΔ jq({ZRjWq+ q )v1JaT6SkFdžX;L($R*}i2J '40SBxꘁ2Q)T>ުssK I9Dڃ *؛}jO\w]!/r;Xu7Q3v!Wf@@[wh5 @DhD͕ KkP@kW}# e5Bd jλ|zwie=SlNRb` FU&g&M&lj SquR+v"w2_# _6"Cp2j8] aA3MBp_% BVڒBdl1S^$`|5OYm=U^μR4^Pal˂.]|*?F5I4nuhFWI׵pFlA :(Z}ī篔`ÒO8sO9ƲIsSɮm*ʕxcmXyU53"N& |ey*-^}Aw^mu\^fw)ͻ cy>V6캐}CD)\\*Ϭi;+`h0VŕU$KJu+(Ɓe:1sY`"ܽkR4ʠa3؏kbŲFݍJ1lNCxMevݤu=v/_-#QWdAQfV[b!J-`dO,#Vn^$Eu/.W%Oɪ! t7]G 0LO-vJ6=O [=[AYnO:>l D/(f)s xuN G!ܵa,ap~(){e16GB6͇#8Ykrq`Z6 &jh6/,]Kn>@ku.b',uxE/w<&G}+SK'e]`ia AHe_V'fz 0gz8_!`Czص`oyY/35iWVK/!Z񥈱KJ(ځlS6N%*|ֻy_t]mz*(oY{e|G([HSkrd= F!^R+ U\.TŎTyߢg8ux.^vLާIZd0Z<\ JV!V\KuěžG<Q[(f(LX A'dgv @Z:( ,/Z9o0@&7 X?].Q¶ f7 f`tVϸI`(pIz`hlpPT]Um vv"*A8LXR@?K|tWGrP޸&}}I )|b/s Iݴ~E%wNΧ9Q]ȝmTq:.pm 8 `]Gq?| OCfy6jRi.Bj2g'L041P)½Ў栬?vJNbyQVQμpxLQ.wSw1O ~ys: #E5_Zŏxh&8u:Á*ǕזLc[X;GZ2T;1]5vF (ֺl̾ϯA_luo+RvK"Y돜grH0۵_!v>:>7+ob(K F`yCG$z6Xy""zae )**7k`W;.^.%Xc/ `%i?pu;Sl ()l"'&Q}f6f짷!=[?N!Z1|Ƿ޾ry622=lJI]RP\v˵ 게p.\BP=jʠ>֧(;,em1L(FZ=[)A`19'a W1`q%X01!H2 9V5]A QX!+g263C& xɔ:f`jH+T7]GV{>V|n8_/k0VOvgCufu5iO泠K.}hY )%4t\|vȗ' : ue燅yjҸ&F ?@_zрˠ˺m|n =6z "M.CF]hC&h;Ae_731?=ƍx}Ub;Vuy"FvÃyI& $׽: <5FTm^ KݭpHk%n{%S/35("M_3cMI=K䙞=& 60MWuƳ{Mv|}];@5Jzvfiٽ&VMy_b0Drn?՞U|pgf9 XDrKްy}C zM }âr3E T0z>`Q7/S¥4XXݹ,$sdŝg)fy6ʇ;`TK/(fzAPtMU)x߶MQl_ڕw4 ug#Me׳#M "I]Sdz^laGW ]Q뚩0ߟv[C)ZmshcȄn\2/"ʛz4_\q䦋|ؿ8xzXsz'c`,-@Z׼,맫Ҥ׫>jJF\!:]Q $ ur)꣖qQ'>j1F\)"T:p[QVעڨKU5eJW!Ҫ#+urʣf ~m5/ȭ1-[rϣRx[(h`mw;r|gB;s>~ XԽv~jT ʰN;nO"M[nJj:8Y4$n FݡvKŠ 뤾c֝5Ҕ3MmRR!G΢)<= -2}[wHSٞvK^TR!G΢I<,dwR1(:Gu45ޟvCExJnؙbPuR߱v.#TwK@XR!G΢I\a/&Y>_* rʜ0'BaiO <9e^z}@Pi}a/o.~Hp;q(63& 3M4'x‰tDrGfw%W{!"\޽y{3 V`_󻿏 7,G~~0gc <1ӣ 7Q.[j:WW3Wfߔ_of3? ;çG" =F_~|[0n?b:hUry`Hl+Dˌ>wÕld4:@G*d)L -C)NG:KBǨ/FceuݲN_D_h G(hiz} 7ǹ8ޏ N_/c$@ U Ïy#+߾r*?ng+6:h=C7HI94H3;ys`‚,i>?0@NiG9̢+B 2eiIA J[µWV/W%L鑌ܬWl~{6$'z7vֱGa> `PѶGTQ޽y?^?SRăN&7v4Kw~^%ɨ~6 s0nya0׾-ͻa R.Nnotٗ&WIpd%\xpZKͷəCëVa3iJBKƙ; 퉱,!^vwq޳ƅ7p^ռ0vX)%1 مgE WQrrƶIM@='>{f)Ef"2{7L@\.tӚ̽m0]гp_φ*ƿ^o`*[4Dߌ!5QO>_{lwѧ*!l}i47q71"yᚋzJ WJ;c4WRR*匠aV^#^w˟1wu˓;aUn{W]]u;c/*?"/=XdY ᒝ7KdԆ_"Um{i`vyeИ/꒥޼4IZ/77ez&F8}N]BԻOUMZFu]V!Ǝ̵4xJѥ0k:^\(w 3edDZI8acgppRB)jc(icxj ?Wx iIFʋpF=gGѼǜEv6P.BƍsÊxIZȎ\U3K{,*U;YiI%wyط2ޅZ5Ha ""X8>6jj5AqD#.MPA"Y ($=5G30 $R=lq"b]5B\SjQGfL QDqCEpe9&,#He;L(4PB{K1[J-$ KՄ>'b#[{Xh94`6v`z`i.3ElIl% kE>QU', 8m)B/_YsM<LW2$^>u+xl3.˝^N/s_o`ȔJ9) pY, ˈάPLsYD4u:'ΥCexvؔĆ' maܗ \BH #V; v+|4f[x}.9 td衑F]CRW$F־a)'6P ΀7^N=i5Tr[z`p<Dpqfj6Lv+z'TS"a)A҆%6`qڤUnZRc^|0nOW'>ObΞ5<$n&KsdNd腲 `3aO #{ WkqB#/ZQs$IT}PYpU<f|vv<b[іhZRCuF{Õ!H,XYlf+3TiJ &U?@+G|n!*ϣ)(R 4DyegĨ,W0d)QDXO(2)QF밴apq1f!}(ChC;c>p2)B@#vĂ"3Q@u[Lqb+$C7@BɌẔ*T!3:bZyW` zn8x ~F2QHJזs$*16 *0Qs^B?9ˆPjo~!;U lo޽`Mt`o-y@ ;K½cвՈ_oG/ESv9gz$&$Eq6TK^*1FpQO\UZR] Y fV|]^X"߅هN̦v 7I<;fuXLM!ӻC.#@{cz5DbZ6>7p!"ګ̋?bMphl8UL WI#lzu~ I 1ҀOz`;"s{)#ۜmySA(Y9rWJET1#[LZ@S!U 8'1!uZR̋aԨ H$-_LfY)$5W^ft>} ݄M݄M1t|FeVƦ))%DJùϨ"iFA_8RBM4uRL3`6g5n-SSu|_`Kbj5;;_X4s?^™4zʒk} ˊ;o/"@)*%<^܎)z*[Ek%^gsP@f FM0BnrS4BZ[:E#Rg1q!VIxaSƎ׌d X:vՅ*ioGdKƻX ]ڹ^<5 ?/'m/^^]iJxQ&t6+K~y>7wOçG!X3yq9aOK,GRoZE*>Ƨk6U%뫫ʍ&Xɑl;)y#WKK_5N\->y9s{t`E¸ee+@FАY>l+L0c#Ĥώ) t]BO/Jc{oS4ؔ)QlU"]٩_.|a&UKIx#MTA Z/QN.I$0F^I 4.8h%B+$ȢR!! -#^;D@+T¸w*8%#8QJ9B @Uj^ɭA^h}1#MD^LzMnDҮphlᲦ'@(늠ef$;4e&>H=YQ<-u,QGUx T8-ٰ! Sd';SAb[wfp%ZdJ((sjlt=VG#BIxGn 2R + Y!4ڽFUM@4I?ea7sr,ڞTVFgmxWaG)`ѤN6{!`zWacIFw!>`5k~{Y5PEsZN 4laӪ"ɝls *eS:M3i(C$%aHKFBXf*/C,EQe`l=ߤgg`$1I 3' d +©Ƌ7ێļ0[*Z@|FZL(?P]VNo'BG`O~8ϏSc_1is'?2Ojd}*;G`]3Th.ۤya6mؾS"e/ hHq گ4h9G8NYhʜeɀ?*Eu Ry KU=kUf$Cq\lYoSA$1M?X9 bjZb}aKRQ f,K&I a[cfZ |̋0AOFcOx)3FIt ?h0$.QiaxL2ɞ5VZ7+}2c"Y\p]H._r'FMA-_Njv;uV=?nV17/->:20pLDz6pxM6y\ zJKԨ8.B[]e/EXmcN K'":Gp/%F"~L3@]Iľ{]U.v/A'EQ` !0(`H-k1?SMg$RY5 (pLYH8,1C11{M.B>iM+W/\ޜ]/6R %_>ru1ht<|y*_Ż;iCm-G Ӝ`)v+^Ӧyeɨi}ȳ2x|6Z) Ǧ޸^̓ҰN YD5m%̼qB_e2v6uIݲev6짝H&[m7.;5ɏ>Fjnn"F+}7!Zw?[bES$ ;|VAGIgЏ~/^PN =ݙbGq"j]}7OSOx#T 1_`(A& ܎Y&j`CMȁ@aؐǨ#dϤ*'i[}/[4Z5|v[4^Z_JZ7b&gӘ 3|Omi8y5O㴝!ΦAAoNk?&Nd@C,( l@/FkߏʴGC_zcfT{2#jܼu?}nhlmN9sW~؏KOFð|؏',#̎M.Fr2%t r K "g)|\R1^yBTA_"E5kjJ3O2@́ C lj>BP[2{ !f]R`:u)0 I*M! +KR FLWiw>bEC bv,VBAH$a"#S"eòI"/UFBB Y+Q[mBL5!ձ$؉h}yרּdUH#l},E @.; ny`uo H TᤦIitǎKg&2`'G;nj'cwV$&cVÆȞLn~]##L44E5 {/S&CǛlo&vӘ5Le\2qV _T8.I _ r42L+c&. ]z6a#4iUBӂޯ>{\"N5~,OcA1->rYfl@W^i+d (I&,#,( ltZ3 P}gm:27b?Az(thu ڗ,W6i1&jh6+1WWU;ڿd {e,%RJ~TR+RW$eYvboҪk)&[p c:^aEБg,ڧY!`W Qdͮ5 8^Zekd}LBOɬЭʬk6 :>0ψ$ۄٯ:gD?yFw !9rSz ^ *t<8l]ŦGzLZeu^LF렃tj1d獕Yjߋ ~"jJU)QF4cVŰdb=fmt*  &hY|a3bf%a DNG5DasZIDdR83R/7] :V|SSH"eJc!@Kx0c'یDpB ٠CcP$!-%Rbm6ѮaTs[C[0۩TԽt-)i!H̿T/yip2N2"VJM@bƌ$H:0BE `$"-nɲ."mrc[:]PRj18eC@%,C&+sMcTf*unF1 fGLckH5 yYZW~6)}2^Jk~/KWE^ѸK]ooyf58kJ W]]^OZ1^B'|~ZMx{2VIr=jixH?|Q2(DZωFjL8'broym秐l%vyU+pɲ[xQG2}ܞoқ.Ncݳ4G:=!L7\@Ku8%v|8[P(riʝ;Oپ$Bs"Ya]0`V"_HwNeDTB(}Ѡ+C_hvvPFzHT@W#Qxumvx%DB+IRY^AQ!bBY n-\ȊG1GcL+igqnΫ!9xH=vN;R449#`Q ^j1ѨQ`eYCHDrR4-vӘ.3ɣYfjg1l qBR*VrcMo,|4u~kaA3>-9A*P(iuE8GOq3}ݣL";º6YW"HQQ^w|=ml]^8t Y ӛ>NԞ:G ɮ_qHll]Gx3ˣ˛wyκ|~RDYr ??Ł ֌w_Ԃ\nk^F#Ǘe.7C@<{xmEťc[7ܛ[rS}4-!5X`(V֎S4Ղ$օƑ!_ߜ1%99iЂP5% #Ѐ%hK{;W5]d#,Za} B&X%Į5I\;a.><,5OXZUKJEzPYoKJl43M˜v!wz͝f ` ^a=]w= !]N5Zk88s~' K[c?n n.H~ns7 u\xw^a GZIY:J"%G408نS _x Cj-RUן4}iHݗbkǫ6%`O/ [[C`ZEğ`f?׋1=Ȩ|3ϳʱP`enֲq@27+6ͻ3P9ji?'OQHl> {%I3,5 ڱYGֳCCqNJ%;O yl{_t` ]H)w0= UJeu%Fz#L!dqW5 tcl.`e=$!JD`A -PZ KQ`̛Փ&;.ݜh,Wܜ@Hhhk,$0aN6::R%pf"Y:ctud:*WjpZ&k_%o4nA?8JYߌ+6ޜ:WWt26_-}xq5¶k$.à\gk/W.n leGKp&H*\X lHOƅ$)]лR'bÐ0r`PpBZkK[傌i] W7SQwn kϜYێT[y>'Ro1+s}=}wqΆY|c*zU,ơeI_}dP8IHy6_? `8S[ڧλG IScSS4TD9@iEݻ3IO{'|es>ߍ΋!d]!,Xv#iۤF=9tmt_ʭn|Tnwݑvm Zp#v]KUmŹ:id3su_ЅW[0umţפd%2iA,|J4`FR"haTP4Zg;7 E!8K)X4+)>(Iy5Q2+P["{rjā.Nw`2Nu >KdKG 7}'Ϝ8Bw7(eۆIt:mzo2y)cN! 4=vsARM؛vOZ~"iE򋦖7;aޤGAJ/JGg?y KksW,Jūv]kڑyĔ>Q91'ɨ+W !)șIα{ɵG[*R\xTE:`*A*6T1`T}/g'BYz<5(4Ղ r H׎eҋmdHfS͈KNU)BkS BZU^Iģ -*HSMTRdC"JIhWLzhb]ޮחE)ֽ6s|d "Ⱟt7{˪h3Dl0DG{_~wˠYUY|+KUrRTbOvw@CJHNܴFNGW'NZq@["fa@6e@@zJr)!bVZ2e<qTo uɭuxtBf=tLaʍÔzE- e<7AeaKeYȏQ!.]!B(Đ E7Έ,Q "Tv\h)T\ e]\Kl7<oV< GÑ(BhUw!B ҽto/ ;gY_-H7FOZ0*P zx6biW )maanѤl+Kvh)ۤdJj6!f+lBrQyVeՠ~g-@5:j?ZƵۏfEJB}cnNyN8K) 1s\d^~_{>]]AZ}{?Đpt]|՞`@U6bBѓ L٣ԾqL3t؃Z'p5\}2#?~eec7Fl,Cs i<"=G[ /]C{Փi48p7CQX.SP#|11-刱7Gs2,,ϱVQɔKEUuШP eEλ2^^U(YU&uKB G2] ΗhO[g fdM37\@d\zyWdO>*zu.cˉ DIJ='f]&2Wp g~Qf˷-\ohS'+CYciwfyqG.r \n4 X_Wlβ *G|x}u<:/?|?c#X *Vy*У)?Z>:H e,E܍/v[y2F%7H)ED8U 1ꣷrLynK.x.45Y5[qovB)|)^hl 1l;tÝԾr~sZa\4"bq|~@a2aZ(ޯ6wέ9ݬE)823~bG__:%o?PN$~N{m6N ܖQ],?35|^yôlhvΝg')]W- 7q@H:=vk9:kɴzhChffLj[()"@'N}[*^C3*+OT(>V Sl ڌD)| dj$*NYQfF{O0BHqh¡%r/oh >⋵d q6?>]|x*'*4La# QJĩ=* UY?lͲ-yۅO7%w_|6>FMy\V\t2>\ ^Y\p֊N>R]9䋾};䷀셾ˑ|H}?\9^kGOW޲'gWir(8ۆ!퇇$F{Lƍ9]^1ݿ9]lE07 O#~KhCzRfW)e F!-_88lly@%%1j7̙)#?KDQL0RQ0siw򣽺S-2_].C- y&:OO\G"+]']z7I&uc~ìs2 A]N|߮ŜBBTH׋^"mņKB.2V@Xt`X^U^健sC UDti"Ľ5ZuNZu7k1"]2ijۥ?+RW螵]j^twǝ,e@^JAEΑ iFkd6 ]YDVЍ v6vl\ј%7,gvsk#e3(PV\\J˜TRŅ+@XBΘWqgFM!^&s^Wg xYby7BqQ(wӍbvoSױ9`L.œŷ'qlj?=0zk=}>3bCOSd~{ dx+G7p7ق<^_G` Q?=Xj!lrmf?^\=\85Eל2L?) lr`௏77QIMJ%3Z wӹoS3L'(" k2{vFijno?6PS! wOᓫ=[M)89S6RнSͤhO˶~׌ GQ;Eʎg%aA VyAree`\i8yGӽȄT' *_~]%yChK^NZҟD+knHݝʺ K3XóF$V=jdX@ >$`WQGe~EaKv evu$GMݩs>\s89Q/+74խ2%=chzq/&SC'@> 99- T O#1oqjALm}5xwJU=$; ܭ |(0#F ֌q^[0fml- 2 ls>HہX߮AR%DsJ.#iNA95%E/D/2*\%k^$זZ" URP5#NJ0IQ{1yEuM}ʵk"@>g5epϐ7Ift"ܢn FNU2ӂk npBU2/Ԫ%<2) bpRWQ)*ŕeݝ䔴ԀM4r)KЄ񒫤kQbJwRs'#,7$yZhs +Kh`W+ Zs) jPI\p'&0xbP2NG # ~E@s,Ўkn[7+Q;19jC)HGV%D(QyO>piBU Q0V|$cb"2>6Tb +~8Rw ֬g oW=?A+M*+Rk<ܑ(o76G$ei_i;ߠKڰnkV蓆6S|<3(%_RS<ۄsjqvɧ$Asp*OקLr K>rKyt0D̫~|1b=c"DdR\ |+!cI:P cH |hD /C.8I jAѵb(J^=B)5g$_QZߠJ2-)sŒ !b@!urD:#RImNCsOLLP*S2A  ռ D%B\di$'T(!W$Tap]%* DL@ |g*TSEhZZo_5Z,o1lKh󊽻=^T?gţ07幵Ӌ'rNCdf;Nc/*\!r 3C* Z^tpeGI.>~X)d֐oj픜mecq/<) y8]ūQz0v2__ǿG$Z '3.& < "zT`lN4tϢ^/#jm#ކih`$ n@*Ft(1lֲvPS2"6 Łl.yu{ *^"h""Emr4I©=5ual EaM@-@,}Sh5a@ծFhoþL&(_v l ݫYVͳgY5Y1sbVYoS$F%W%V'ɐJ,1k;P+?^sB*tNS,_C"0* 3dyɜ1XE xeL.y^_)ԌtS`A?(sgx1;2E˵BVCe{B@7J%FvQڌ ޳e@;^-I'̃Ѣ(":a" wekwT&РsSϜRߵAg wmֶ梑(,jDR:}(•8B&g"% E"L<SFp]9YPy-Ic/qrZG7n3i5{w LjPTl͡^tD@Kh* 5T:}^%;k٣ ۼ[$oV vȃX**My"9 qꜲ-2;)+=I`}*G){8-P*Z]Nw5׿@2B{Y8x9;Y{y.J:V"%^jVGePGq(v"&»Zyg Bpmըr"!<i" H}pr&廄T$ 'X i"F,Hy^oprAR#2 u#8nxK@F8ZTpp',ZDlLN× @^fڊ,s9g>wU||^q٤#Px6%DOBDy T2m|PN$n5ů<RB x 4fiRU۳1TPt]v Nqj )ɠ%ЩMf,t#7 02T15TԜ7Zk!1 #0~ A E5p4:o Q ceVՇ[W5H- =EYhFPJ:1G) pG_]luҶYmE;֤i{4οd-ue޼Z>4Bu0_1\'o > ~qk$ a-.3D;m{{wْ5-2἗/5o>hڊjb!vn+*r֚IZ+.h!RrFՕ=:Y`僤pt`g' kƸO"b-qX2̳xbHXJ˽F- -- @.ZXʗ+ W, J\֥d4gJR~J Wk& zn<؄&QL֮}ڟ ߀2kXM/˳gY/ϲ^_Mh^@D$%(p^F *Emj_v͕n֭k`NUіKBDVJ+:77law!KB +b&C ]!.Ov|< *rIH֐lWq8gݥw>];[(Q \8N @ #S5c<8[pt WsT8lQrпgEN_2-?Kp!ɶ:,bR$a dk(PZ E zN6M#5 BMgEO9c,xRh*Q|[ioSĔ͉[&iлڄA?-mȾYF!Eeajb4UX=\hUCù9>dئoy 4V <9[?K^ۣ?\N'_"_UWE 4VLMf;iTw~CO~]O.1.͇x}}}a68aGכr<#/Nl60P5F9bя6YjrڞoKʖ+kTB=ݔ7ߠ8)-?79>spJ4̬;m&GOoޞ'ΔNWq+6~뛟oP/ލƧoOq>>%ǟo_wϟ_g99)hsFWg!_Tpo.*z~~盙Ry+ DaȆgʛ6Wa鏷/ jC*'QR[<䥶Y՜6״T^ҀH*%z{z{nʫ=e!W+)<{5ŧXݥpͧq!Ls-kl=@ǖ,Zgn׊dF5 ȿM/ם5[&ًp1|`?,lV^ 3ͰG7y?]-΂x>\Ve䍦p`>_=~T#=Yw ƅ߿Rcro?a20ޫy/aOl1HǴ@Xr(^&Oޏ&k<{5 Yp;/~ `| #*,!BߨP+ыћ˻L/Z#59ZPYcߖ,bo.GyQKt}o+SX j-Wnfǰ4_194x~q gx<7U__p/F 盁x}q2x}q` H vyj%/yXTYA9̸Xf?a&W3VdCdVĊBB|vX8pLmQs1ףk/zv}PBI 뗲gǫ)h2jŸ[!4[θ'Rkl8B!/5BJП:5#*!JS:M2B VH S2P7ĒK̼>_楴"nq!$gy0|7I$An@ ޔ= ,*[NG͝Kho6~XJ Il~&q7(~DK=3}}N2 v%O>%n=ݬTNc&B^L72x"-a"ݰ9Scma45x-ݴ}ۺuXN!k';p3 W{n\ť mQh=&xpX}ta&J%;W%fSxG]k[c)Äf`HQwGWop}r8G=ʼnGtp1Et[tDIp2{jxt}Z1/5ŀɝ'c1נ0o,ޛI ^382oirsldÚ7]]^n>Y |` :)܊_υ),@?ɞp Moh"i?,ZaWDPV[uT(Iэ]%)+e.n\TbhY9?_#$6yҞR,?%{a_Αe4֛4T3 vj(X$nҖE&5ngC%iH b%9V}t)&[[g66^Ynwt)np))sG_*1FmJ /l9B@ jՒaα:‘5F#ɵsp=7%.k v(../ 7Mf59S((XOS;F]z "CV0$0XKe>BcFl;ZڴKyfL*t,,Nvʎ2];psZe;8M&/JA+/)8'sepE1*dxDn_JMhsk´ >eCJl)d$CJ;)CJ;S!%V[9(tHgXQ_ S0 0;p(@aԞHqVxOACdH#nʨyܲ| a&:!n$3G%0d!NJ$8&(@= g[ V+-m:?vIK8Nߪ"*.6"8ߓ8{oong^黛Y#Oݬ{8+}| ́˺S&~9[0 JwK >@=r x#pOJQc,ϴeS(fpE_k}V6F]ȼ `К$`A4,.NK}ӵ]Jt9ehN3T*kkdX4-.p ! gsj9p IXNrV" V~e42K*nqg*bꛫj$?L'Az~WͥFͻw5ﶬy`!) (T }PEhlDTM,DY*M8 qU?@v=S]^Ll>șyu{T޺͞P[dBF-^OJ@)Iu9UESH*}΂ [ Euهo Arh$(-#}U2n75@<H!Nӝ۸΢ĐAf0A}KAV|w=L\aacau:DB8V5v9@2=Ւ3 WqQqd6p?R1{ScAcyA>8GṄ0puHVCE"d)\fRxEe]aC` ռuГ)I8;:>uq:xrU=ID{qJDza١_zT<1n~TMEZM"qN%6nTF։J2Q3 +A T_Ubɀo'j2a2q& Z 8τ@2 ~`N#)m'/ |8+lY1D!ă(VP@voP֌E>/XmKEP%AQF1YlV84x!,N55č$7$x)5‰A! 16Z8+8@Cqx/ yS並meWN@ "nժ_ѱhNQ;J9HÊJ0*#XL35͊G.Ode\&<t1p 2E{ ~C_t}3zNeJ o M:S~ji:6fC;kIC9MJ9q>yNg"]Ȱ(ybyqRO!eDŽVmTxX+[ (fkK!bDAtM5p8TW$)}&6gYxy~PZcTx}N0f<6@H:QXCFNG[NĴƴ$fjV+$1D{Ǽ3b8.pxLR#4(XMfu%p "6+:_pWteS xl*ʦl*ʦSV6UZaM@WtM%JE ~pᇉ%9s;wBtpw^wx_o8)oK|7\eFԓzǓ)j!qb}#!HGrPK 47r_RwlO*e˾$_`c,Q^\߯1C!)b8R+[^bk4~X5D+o3ǛI\\O@F U Bt-5".FMQiFMMôr`,<$`0C}L5}vγ!m0m kȆ=zU,Z4q8L'mxmBqR}f?>y8Hp wb!>]rgD E1Ʋ'/Gǫ| Ac{ObG)|,7+߿!]ʝ=WM%?U8(shw~kIzbuHN9َφOE<_NAWGӅ+cљ.I^[4a\n7iJJlrz2?Q !} ddHS=d@Ш7w?{9n^Cm]7+.XZx?Uo#8(|mw[_k},uۍ-| cꊬ3ïu[PXP{t*bajCtnD ΒHZAe7nE"t(fer@oiP2%bOH}*B@2T *3ZLkfC?_}J$͏@"J_;i<1Kݫ>w]XUBXqD BUY zFQΛHbk }%LpRS ɟ"hW}ŀK2)Rh>^F[OHk~w"SOQ?EATav ]p9AS%h5Jr9+@Cʝ*r4H'3 һ.UzmL( &]_t9J$'T5E<_.2$Өe7:%hŔh] տ2`8`µ*cZ5 c,B5cuZxmV^Q `Έ:FϧP4HfO5IYSbcBIFY}rSD؍j(U=ϥTjU)^1Ƭ1^#QS;F!TI121fOHN 2&CBC(˶VS1-b+&KY9E-Lɖz Z{.yI]Qy]n௨l#2)͔2r>6+gh1E!EΙ;Jy9FO;MF]'1C5$f]+8Um1!iMy‮0e<Z2GS kc8CjM"| gl!kj1fV:(xpᇋZ=KBS@sS!#Zb))i65osaܰ]g$DxԹs 5 mY{*&{e=,w̡CC<ԄW zn]t$2'{_s\P-'(ǽ2osF`VT_PԪ M񨀻(Nx󧠈napã!eъ48LZ,(u‘))B eNh7VR8Jn7M77e({1Rqu#bŬt[#~C ݶ8c_f p~jHV/q8 ;2 c˼_ ݅܄Hlc.^LBo_l&n&d g;(Ԃ\R޴'eU .)2NKtEC[2vF-*h= =i 5-9@MMJH#o͗Դ@|ם)+*)(h 0U9G/D\<\K[Eʔ!tPlъCyGi?@kU!E%7R9iMDl. <%y3sRlڰQ;AM4m׊+ISZo֪T>(o^*q PNk.QnpVX8;ٯU@E2eV~J:*n+OװKr/- ̗*M*u܈:ϸ_3;ܽުV3Ly[X\d7[r?̲ˋЏ%Y"B:t}-0 ^7 (F D ߮NzQ-%/ۇ8i6~Yg{¬V戕[15v)Tczu/ۡe _`m\ٽzztO-ԉfs؜++Mꥼ[PB߭JF0eW9KNyܹ+% 4 "kCWÀ #,X.irPK_1 ޼ T`L(BNΥ*"tG#WԲ7*g-9k5Ac_")eQ[y'OˌUA.޺'s7SbvwOEy% Y6̃F۸ Z]߱s/cDѪS-)^+zu~{K ]mg?A܏q"Sߩ78֕)m񬯈Ou-Kn}H.2j9ǧjqnh7EjׂGn]y":]F&i΄j>$䍋hCdLɩf"1nKe"\ ?meB.G>:0ZDہݲ%aI#`$PI؄>HV ]@YsJ?Dg#.!J#j>vM!hkU󺿮Yv y}n37T.%{)@ZpMPv OW ) 0 $3|GP獽8Tg$"̚ಂXP/h9F}̬E {]y^n܇x$wM`kX"i#3^,}8PsI>O^Hb6@R:SS0$x*Vħ@c9qY `5TXB6zc!YH \L%;yڳ Fiq {-e!96!uN,ĕBd%/Rf V"P̓\ &׆a66Tj+(7WĻS;%KM ^er5RċBB[C A@/:TjC۽1S\άҋR!f !w*j(*LnT#lj\Oe5B9(ƙZomD RSxdzZvW=h-^bZO]ӷa(9J%I]-~d 4[mX?=e]X hhH'+<쫫c#x~]XιG8IF(GK9hЅтє.xN)2tQǎagUǍeLEYǤwuܘKulH BxuZCvr\@IChn%˭}=nF_r[ *bƒΐW~ P $Ҥ^tgUc2_up#wp(q6T^m*ch2ߙ.՛"<,PK{+oW% ;W}JM%6eϴC>>HQ DZ*_4#_懍iB- =-"R17afVlZ>G%h.z-2`:$/T{P'C6K'CUNKK"5aJAJ>RIR..VnIr&8I0scs`8&c] cS+{R}~xxsˋH4re"*pYv[ M}Y~h>TU Qݤ;WUK3R13$ArVZ{xoHҬ<3_9#\i6pnY@UU{egx'hhę @Avr?nspL** B/\ *BLs)r gF|Cz8I i ixD*|OOCIzЁNR!`hZ_M)i.^: I,BňkED)j9HAaM6 4n1~'ΐm[F^L;tH,y74](tDdB2U2x{JXy>A RQB5PwMB۳s5[UlG[:S3<q$1*) r٠{{/jZ5ɗu6.Fg̖ӭTeYYͯu_3[NKK>,Kv"0 1u(^W?ϊaU @StxG򫋀oxoJn|\i*Y?b[Va;֋Ux Y8';pTCO2;UH.9"8sPW*9\PhkmhER^gw C6- 'EӾA$Ltb[$I3KR2uER"-&g LSrj@J35pӷ+Wu6ùA+">|d<~~6iƮ|7o\)LbZzwv-[ Phd:GN\[ek1F/D[kmNGI8F?&7;|-܋lw8G]lxn`Yo>nO^: %#}>O&dl2yyw/'AJ$A3=\?+Kgv_OG9cv}_w?\0_2OwsX܅/__U1~D;7'_q⎺N\߽B?DpWڽtv=_|=ϗmwCBE^MW onRx8 g7ٷ/?_vȈ,]}8 BP.QEe~w^Q2ܡ|/_ nn?9:&no^L~G/r"3u8^|qc#>sy2Mg_rW9ҕ=vݎkOQO4K/F|)r7qR|-~C#HdZc9;~ 0Mf )Ϋy=|&+Q#K|=X/'EؕW1/yYK*T+ (@3k!ġͅxTM?LZcPZ-pqIy7N\K6.V{?w,9ߞe۳gQ"ޞ<ޞ o}INbi\fgfRS%7vYAPR Lmd)Ǹ( )DREDZFHp%.*6"$ <6$Q:$& K$"a(Ƒ%(Kql\? t$ZǨ\(q—HX81Р2MTH*VT'UE (ڻw}s7M ZrNo:dsonoZjMsxkrp"V'eJ"{_o<}\HzJ./Ϳܳ'}i{ۭATQuhi#?#MR@L$0!!:$VBL6[jpBlj9/Aa-l?ahZztY=^]vBmcۅς,,xݼB+B5&1ZQSU=r BIUAVZJn7ːwέz2v= pH~`3}965gV`im{XTtk  H3 ^˲yy5 =ŠJIeFӧ^ᬙisKPm{} f6'S&&{Wzm:K^. P][ pY*l[H_&΂U=K _5 ;>B-K;"kFOU|ƿ#nKk 6;Sݾ ~F 1ZU^g*R$od5ȤjYg#+##g%%z,eW6.LIf5 $4 0j)*+^iB[i9m3CǖG[Kt,I3ה?Kʥ;I\ýdY 'Iř?Ya8۫ |؞]'7;i_3QWLĖӴw@Ѧ;ocMw>?M -;?=G2Mɾoܒ^m#=!]}Ug=<QZ}3 Ek ˴.&ȟOH$j F$aTJDkW%$6R!UqȸHDP1ȕ)TTB 6em0dDduJ2&KHViČ$N ZҹN wtHSUfNMjv*aAGE7krm[7@.r,}FGPJ 8{6J)Ɩmgzͬt^/}ZUWu6CK6BEDPGՏ;³ }ԝV܃WVp@.muzZqeTJu A{yRywxyJLG]Ƴs~fqM\g/nk5~|wy~w \OFCg̫pTDIJj#tκ˻?C̽*}sm9'9hzPۭCkCn_SZقh@ . _ ݰ&, !:u+WiI5U8xB/W5q {t|΁Z؛[ Z6wD$w6eyqػ!_ {wyu;.vĽlוG?U*aԁL+unH)wN9SּL/H/gncݎ)U?³K]҅gx@< 0R"x50G`V Do Ym aI77+6Y" ٮ/lIO[0|:fM?$f~ΒtږtQ=uZa5\ݯ'ǫ=+$7U[O.U345b}-2 gmtݓ6> &h$HyL slϸuy8MTxD҉K'e/ޱfVb"Lb*Ih D&%6941"N6Vh+w]E/Y"EŠu] ʫ B_M"!E2K "C2T(*,41*CWԃЈvX !(Y/̝LP&j>?PpV_i*o<.VwjSJpg2*Ue<P%)_  B"6"ҘDsb)Zd[Q%ڔi@,"F-\hqe ;[ K:Iz -Ez!` "Sj-5E+J `H7XR,?O>%JTے9SAYC55T\͒x>TjbJ(JhKT-ӪI旓(YjO9m$JO'egbmDӥK6rhAy%2)ݩW0&-#~("Q:aVkR*, &AiJ'BIc%:4F(! fZАY( ZKD:CQQ#d!UiqâZZ?Cͺwц3n9T 8/pТRj5Trl<J$@d#nz-mfFZC7NgT5'v[[I[R_4kWZגa:K[|*ްJ Z\%fC@^Nq7$ٴgϘ7G{?,~ 04 n,Y؄wwd<9*rg_vGN,W RKZ1.5ZU =`Z]Zl@$8Ns nsMڣ[S.wmXA1pkbeTO$$m;tT$  &ZwTAE~w\OFSA?GI4Ttj*zЕT<t㾱*iC=Z#е/\WSD n~1q"_)Je)LP'M&û4r+ yVJ* 9:BkUD-]٠ٸ\:7]ޕq$_!1&ofpG `clWQKpon`0>d/ť,ªfYڭfgE\̻~ّO71N*;TdkUQɳ9W)*Gpq:BZHcIJ؋–trWp*[W|\q{k9~\Zc y/*نO*J cPcwq ?.ؐS3 愤Dcbc.cT3HD,202ɂgl_ |ҫ: iH֞"_ӧُg!\;ZfsUK a{}jTEP5 If{Hh`a ca PM= Qz/V5w#o9ip2PwYJzq/daFa!uH X"GЁ"(;1Nv4.!Foh`o!%`TVcR:)+udDL*(6,Fhxe@M82Y: 0K0(%c o:#Ie QԘ!A`6aK¹_qR 8f$`6-L (נEteҲVN6ݼO ډ.7|#!V\-j38f5w;,~k$$B h mEQ1S*<;LA!5&g&̕VaC y+ϵ6()zo|lXKQIa.boݱg@T҅ o#XI6#T"Gs rXSn 6;i62)4OTɳVn|6,'}, pQ վ>p4p=.mrEmˉIjڒp 3iQ1](t >ۺ%>Wغ^?Mxbâ=mPW4! kx[t -F-4Rn ''ZyY+,2*,jEAUGgooy;5Jx?_ u_)7,S܇M+̽x '2faKDN+bu+!nHIr7F̚je~bT"bg9JIrXerr6*Mt‰YEʈ+-cMљQ] %rݶxx0GY1Elzd^)lsR`ʀAhhV}_xE9eٵ%Y!s-M2c<"1ch,31fd{kRg<`̠N+@2r5׷Ï,t6rt&M/Gm2ۚM@X&F#TrlB92|ô!e?Ƞs2l۽=?G+Z5;{ Bׄm3;Y;́%Uk!A Vv}lg q{7;>׌ranT#о툾THRYGc9}I8y/nڨ{wyiThjgN/m5\U& 5uP$?G0G[^G|v ,#Vn6XՃ]6- }Kes5k(UKTCrQ'ฒec$ M0ntdZJ.6m74R mӛdAJ)$>l] 9> r< vxvʹ,ww]nb`qN?[%o4,sP 4i0MI3Fl>%lp fR}WN'W^BhE{T RVkDGW_ⱶX2ݫs;LvjP./].hk_Ar ,6渇s&vv:[.׋t^us-]-Eͮe%t%3ifM @MQ+A3{l#{(^l[Ae׶M<](M*oYt ,uG8 v & { !Y,v/yP܅yj^=o`_^jHpk%-0aKdc5;snc jx] B^(%pTy:1<ʆ1I+kF܆bpMqa _Zd ?\٫X+G WnD۝VBWY ZK0#Hh{^e3P&O@ڣF#(޻F*QuFW= wV ΕƜ9 !P+}&< JppyǘfϏ٫(;-{zmv 9e"w"H .QWН~f\31'B=lAw,+I[:S:sʨ4gY=}]@[fagGlT!:%\,j, Ɯ9]AJ>tzߓ-eӈ<"$~TAoF?/|$_,hB#|Y,-.]_f,9SM=SMTgkaef In0bcпcqjS&*~#Sm⃪GŇ($X.Ou Ya&N=a&Ny+h#fQMH 3"Iq_di$4bV\AÏbM7gi-b/CWb)2bO 'SN};')&SI &C TI9҈297JJS bUTiU}}@.\ٯyO[+_>0_B/kzFt}&01a_ M6w/5 xw؛\Eg7e>pl^4A){t>ۉLpԴ]m<3 @ie V2wLC)6~דgFmaW 6QxVZmg&i`*apxZYeԁgtw*hW7eXRx m;JB#iqڷ",f-R%ί|ّ6xde"ȝ2]5(IJ6NEtӒREC1&R3.gHLdRFKgmRg%6bX'#IY8x$jaҮ;d1> \ Ϲ|PH .CE޿̼eڛ=p4]@s~H|&.en_I7}}YN|FAf~9>k/?Ջ؄Cr9άc G\[JjnY'+'^@OO7SJ6͢KΙf m|/I9aR)N׿O{ gu4 Y<KdcDLQjr'gl:q;< q ӎ1]n!}=QFӃ:eQDA \>4v_bx?k=w!Zks>|k:є\U5]mX REc@+<9mG\8ik5gLMy8-]itVivO2d-eTJi.6VMpS-)Xoooo7<c[)]"ıM%hmdAe:" QZB'NAW}P%߰_& 󍺜?R3 c 0T&T".#JE"qV W*g2B,["_c+>d_D>xm!X`œ2׼BRv [=CQK\#0"pSjNre6SL5xJIOI_P1ȸ&$6ePuԏ=DħJFJ@ A֒˜e,&PXei[FpSLBdqiD\;Cuek$Zq?TQDa)O$D(ݥoJQq ׊B+ ^%^80  1O^`n+O=!%I""T6*(R fȨ̈́c$Ҍ2P윻)f EԪcvu()$Ρ wpOnͿMS^۫ǀ(L{_E%R .g;kՅ`UQ_ buE6fДE85ϵ~c[zq}4ɛ~LtB|}=|6.ߌͻzz$i){Y og^}^´Wzۿ_ ӴWg`l&PiG÷ѬK{̭n:IѷūO%MFq/HX\Jim :cx.ͱ}6? Ff'_h~?jx 'USxRv}E~jt3}=L\_DF|zP`YsdFqer\ҳJ<`7L m#as̷>L0|`6|D9Mi2bϜA)r Ξ RR𭥤rB TnSZf䢺0;_ZV>r4gItւ3s.%};3xE \ͳ2ǽc P \cO?"HdbWlR@H~Yq"WLL,V2u`eŚ-SrkޱlP0i vwʉ!g杦hKȺN 94v+ǒhed S\p:RTy|ǡ)'X&RxgZqUqd|&>sV-+%nLFwHf$d ح喻j2uO/q-Ycd%F [ tG{Z8!E  t *(T8B2/%q&ZeF5+AT8 e{Q.K̍T3w=ک0y]1 ˙PMq&ƭZj㩅gp=> F89A"4ᇊX K! iZ+xLr7G1|_L^XTvIt@0 !p#f9E4RLjcV仒$! )Mtٮx;j`])h UKw74JQR4R +,u.;L*vV.UnY''Tk졿';k,:4\w -CC PL}(Cn|TsG\ZӊS'U{R%ZEV+m9<%;\s6{>WJܶK%zc2;mK[)nь&~Y/6ӗkǫsko]RTmI m]ybuX\+>XMu?lWvCE {mVb, "-DTVUއ0$t~UhÁ vQY -ʣ>]6]6D : $AH̾ s.bœ)& ٳa}M^˔H,/~71+bՙ"lϝ֯*kooz-f dQğuW#K~}ZD&ߑRn:5h܂Ix:sfO9^N_Pvas>}I DT_ QM}(GB\D MeYn1ȥ[aV !G. o~{{S uRbC0;F`>}vn&<5)\)y+1, k*S 8, Xb"E͉yWC*=;$5Vn 3T) x,Bb6^mx9!|o crZ~B[GD >cf0B AiLx,cċC M)քD -3Ղ,mB墺0۝jh˩vUGwR2눔+D%g jjN3ũV.M(hwVnPq-Zxerr"ZHt~~[n1'r1b`#:EߑG5hp͑M,[r""SRhGi\ da>{"yinlKjy`|{=Mg0Z`f}Y`,ޗPp2'= =L.Fnz;0tڻmyL)~!dڍ)uaSh)iI!NF8ue& )EU~Uک8*3.<0 j}ŅU4̇R'H,);J2@QPͮ ܙl 8-%i޼xv;uD !!G^|g\ɨ?y6KDMi>~1POڧ\ &3Ѯi1AD?|aQ2Kଵo=N`]'gX`c0՛˴vqC1̀}\ T#@_MY OㄒCïIQ-}C_"Ozfy{ Bj—A}2Gcdc|1|02VeHK-$mW˗ e[˾v:1&ÒBҢ3KY1$SjTctsA/1EUӛ ed<|S\Zc]sѓ%zD$.}oN6>%OC ):Y 3bYR)[]&+{!18BOvfʗ+dzH p(WdaϫR`)]>i@3^mxX%QPZ lMP e+j7X=Ucpl`ge'yNz󡝩Zx5VO*ZiQ)%a~R ʂQJF'`>:9.[mZrrkLBdۓ-ʚesKyGP)b6kC1w9PO`!=2p<BST0b?C(8 AS9I@0a>͉vnQp$Їa4/bqPih/Ԇ;2#8D\@#,ffR(ha`|@~p-}?kiKtT gd π{y(ټ.sgvփd3WC]фL)9SrNuߛEz6j02j^|Ɲ? oS3)ƒؽ{ח"%P"W/?<0Mx>p "D_g7 o.[N&{ qz{;MLPA[,~\hN(v͂ϯW #.# CTW[Myf"R0ܘ*鍳F Si|s m#a$〇6*8I6mq Ȉ"ru}`^_/Xb;/a+IYWDvtbeül}+>Xp9MWx9Zˊ$89{P"FuNйbmXH+.km>0T;Z GB\9[n}<:YzYwrBSNp)-Bge܅PFp-8|&QBŕ:HOεw!UMN(=hި h1528.;Ù3B]2>MO"S ʒ7_= )#Ȏba4{Դ՗v3) hk)u}a DYM%=O1[1w@)o‰#[A%BʕG'rcIʴ+td/iDyF):tג+\3Gd>AD:5\l1SHcDokO`?^z`$I"k|2Lm.j%Қq2w^-.>op}/uFꮍ]Jݼ֣6!7k : ~BJ<^)Bp`Dh03ϻ.m]M& F1 ޴-/%FZ {ov;1Em 5Kٚ}2/֛}n/YP(sBwUUuJUcSU(nUuގU9J r˷K4I˪^*lSGq5UZ=PǏg7M:(L~^.0W4x6g4L^%Tb֎TmF<7o/d( axfϊ_Y!moBL}Q B,c(E|p)Kˀs> pd^!ހP{ncQB$<ľ01 F*,ИP"5EZ(@'I(aS&Έ@r&@BJ&szs6qCg"" @ncdFֶ\nY^6)xrbXxY!Xg!ðņpk;4a[`n@=y qNOUBxK,JJ R]hq(҅,Beu\'ņ Ti* `%7PDٕWXRA g`LSJXYsmljq8dJL1cM(L8R(6fUJ ĬRf%&7!ߐg7πʐ{L01L=x%roWXb)dlףWV^SG1 cGrN ǻ@$AP/ځ@'\[_Ä͕89LZߒM"vkraCKC9NKK{na)hb}.5C^m6CwClz3Zd)GnJEgw))rAFiFoGοuCwT{mӇrt;]="?'cz2^?H}F#{Swrjls?gPXY-%(\Քsՙ0jVɥIZK3Tld!_&ٔwp>yӻ1@A>#ƻ 92n@ֆ|&dS'S|_6Gfű7~_9!}$0 8A*Hx{gP+]|] Π:GԈa XA;kj ug9BI M#/3V `#4'cf<Ýэj’% a5}{lѓ˫o|wWO2M[yk_vpMPcqݳS= ~ir 잊(Gl;́t}^OpRݴ 'YRiA۩w Adbtg'q_<^ͦٗ9Vw=Xi3J[ۛj$8isȝB D2e{|q `f;ݔ[=t焑L7N5]ֻ֞?(Ly^}kzWat&xͻl { #-oϻKa'@opSgo=t` E ]#);L˴xsJj^u 6avlս+pϩsJeH46kwr˥+r =JE(SnP-)h![gq²"|,´T`a@ lB CUU(z>_ߐ w0BADq!qjBAZ qaHe)Z*P"6F/03]q <܍]"eqᨼŞۓ7vv~cC7Gq7șY<{o Š "-Hk;f u 8?z*5us1 u}eq+} N2qhY|˃?NjTyO-pbz#H#=iA`wF2kF]0_ec &4Xfc`ȷv7ک"1,J HW(r#]L[bu j z3\:9#Ȍΐ%Plk YBKyFA&0罨 lWFr UcjHSh؊s/4#l6K`̏,r"[c!fiv 9(oadHdhlnXz`tqϽqdH2S֙)b'3Sޞf813Y MO:j'R7$7R{cu 4$."rb,*?E fOKm3]F*/۪_m<1PygV Þ8#b|Bvz [ډԐ9ib1K %_9̖\ݻ.[RFݯ1T; þ:uFlq]ё@'OQT( @#ڳS8 g Wsu R]w)$S!S4#UB-S0A1UG8B])-o%OH9ߞ57]="~-c:}{eai[`a13 `@tZQ"$\j)eAG̨BDTrJPPka\Gc8"5<6`9X On_j"gňH s.Mqx uKd4XDnK jlmn(\ VE.1w1. L,0܈($-9(7DEKvlqe4Aӌšh HED&c$ ҄y(i}6UHv춆S8HaD}%Ж #8݅D Zz)Jo=" x _0<r Ca:Rd~ہ>w G.oKRG.^F?lk~>el_]k4mWnoTm^B2c 8. VC6LGhT`v&;!扪\j^!ֆEy07 ]in_m.nGmهvfY1w\۠sѨfo &`'V ԗIIp,DvA%\s*!UEA.9@Xas8Lk{rW&:EjxHHP~!v,A .ܰ7pFU%Ҟ w<#g%!%/[M/_?0 J8#Xכ9XdU.AƨǶ=Ћu [-J.)Fyy'*N*Hͺ?~yf7cU?9ZbF%p{U-YiՖ 11鯟gEp?Osu .F[<1\+Mv)&JT1-Y(omIB g*7fiEdzjC}vѨMoF2ѽY/9ә(o.9*ͽu9NwN[<%b*8?-O?H}F#{SwrjoWlN^2/%(ݞC9=3a5a6kEu9C{ljBpM)o*uw^ɀ-I}Fwnt%nRޭ Mt&7+":^VksS8Zhb[Ldw =7GG}XL$B>N$'s5!wUPv4U|svֹfׅc =eMq[\6.͍tԴ}5c煋ōyn ?81`BH!I^=2i КEI)DŽS8.(c)K"^iZzs2M~z/][ᵦȃ[Wǐ:z}Mu wQ$A`>Ĩ| kP.GA?܊<}5)٢'ԯ͑K~a}*GE 8(J 91ucT !Ǧ% ە"dQ%5$ŘE 2D$P(b )@睭^lQb&~sy{+uRɇ:lLOyw~Ӛ2Pd93#]o??ftDd/1'3> ~yiQWT"5RW1 `6-NoñR?EmP,ƀcQ26Z)6Vk4nԃv F5آk$ix8QSF|Ҏg[brM״ܓ`YC^>S?u2r;Mƙ:@4S r>p)K}M&@6h{5}zI@!7힞!48=&l w;?i_ `hZc90rF,9 e̒``@^p9^fkSӣą_F.óCڍѧb?̯v>pQ Yԓ+mTV7nP_A}l,2WkQX5ƀ&(JX2$ HMш0CF`2{5ѹ;>voc}KTcqY?L~+2nA,#Ե%e4c \X*~/^Im!JXbZaPQ)Ƹ`B#YJ (*ƒCQJ}PJ}˹_bѨK2\|Y!YOu0/gQ@BڀBJJ5/։ Zi5_JWw]io#+ɼ}1u=,ƁQ X-xsY%˥b-*8hUEg27z~8¿rza[>ti=L׹<"p~gMa|1Mm64oߚ|5r6oVпf(Eꀥ i8&!A>v0Ԧ!eO4ߏ:헽丼h̰ҧ9-K&%ICq1/A$$0"00"qLpel^}s6~w63!,`z"-*JuVI\'Dp*b`GBX6 9A)x8R5Ӊg`ryqQ{0קlX(2WQ>pps+_;õL}  p W;i)|y-T&&UZ٧lUm~>,Ձ zγY0#ppB!R ֚>U3"tf,Sl{;ez`=TH~2ÿT߭~AruT ʅgJrQ [!aۃ*}Tgq^rIATy/}|E>jdG) 6xU&]*q0ş&W;frs>~<F8gkв?F>z%Q_0TNFSD)CB?lkgo}Wذ֔bot/ %-k\fNYgCpiCC(WG|و5 Tz8 }VYftMƃ$D9m06RU8$Q-\ $C'LR!Sg ֣6Emt~:NӔJd\%Z,әO6֞[֞>-BվEAAoݺ,w5qQiXi'ٌQL` ҢQV>lye%ڿD+Δ.غ J( -%.fik/b&z[r?.<|bBTP(&O DQ D;CBoHz'8Ӷڒ^1e(xrR@ bUycmP uce>56Zdv9x.>_n|.xn#9TRF{1RFŭ;rne$ IBBh^* .PDCN j05w}rkfKQi$:[MEL_W2h^KٟK}ezΐ'xijr}!{\Vu*R~x_ͳRӨ"Q{4>DEq՟~kr!HETkkMQu0KO0_jM+ ;-В.lzl-Yzwv.Fb ;dE5GiӃwsxuO?x+&Q`z8C8YPsBxmbb9P^jrb_&֦ ukjX*"xpXFDBG~+> @(I$!T)OM$Sf>Rղ!O,p}Sp k@ntQDU0 Lb:QaX 7Ђ uȗz6')-Wi^w{Xd#RS:ZPFC?P5lX`ќcC#J0˖PRTf6I1TAIpVabBe6 *N0, 1$a +#5cQs ԟүnr)X].N .4vf:nLaa2҂CµCsrl}Ptb+'>vzA<]<6:yV„lFs%V@ض7eCY2xf1]~WoT5.Ȫ(zG%J&€5V%Gp;]T}ltj@'L #dc}BHLy5ޛV/)V-EVMqbaH\d& kU1ӇJ/lH}г~u@d7f|E~  q颇pz`uЄL%i8Tc$L`@X,TC fU E.bPń~aWn?fRuEsjIUѼOcBRh h B9/즦H8J4,7{iF8ˢDۯOB5V B3h̬Z1=b{1P(ıA,Nd5,bi+%X(kֹp%$Q| ed|[x~ooWAmQꃐiA;f{aJQ6dv%H)hW ÞR[]v0N/,Z'a\?0^k> T}Cxk>-G41`̷Dlfn|fpa_C_LGal ڟ Fi{}Ӻ|I 0,HT#!Z ab$5 U| ghσ\K8`ρ ѐ[cQp'%L' ޽58?(n&+ tG!$.#!],X杬zޙ4oݺw)^^\# )@aNֺuJXP >~Ex=nvp4@;67L&Tz˿mwYBE6}[HZ/bqu_%dH>@#u d8I~SqcT:7NMN &b& ✅ F"& 5D& 24&,֡28DI$1ex>grD]{) Y'.3C\˙eX1pElg&_^.W9ưk6G9z|@mumT Uجu1rx$ȾRRpBͰ9NpӜ %hK-͉08Ủ %2!U`F4ݔHm "0 1_XW`E'.R N D7.&=*v|].G石A:< &=8lg4$78 'e5qy8i3#7 ~޷RĿɍ1/-lȄvy9LfӻgkFdXR1 81I( iDS"7`O+;x gatz wW Awwٮr~Sn㯝vSY{4^C;in0ED\b!Q A"Sl={M=@L4)hV縜Ķb;޼<~@IKrD'S[bzs䦘<`sbz7LZ҆񚑍xE뚜o)qqeדDB͕̓;/a1ujzmB9zt3Q42?Wtƚ5k7E,^IH+P*<(Frl2ɉU(%A ưdvSE}yuUup˗ [ʑAEZ 蕬vfu۩q# ~N8 x/D/mC6*NgMNR#p{'hG F8D/1mߦJ%hy2npd{s$=$a*O{y{-( D/GNX)DG ^©q ٮ"=Z٥T S`*ɋҮE)--uϤȋҞE)/o.E$@Io5.&ņj9Q_)ZSS9h޾UHx]W!q6pk˞Ol ڗvaQuċf(4dEQx'£t@T:ZvyƯ|+ ufW9e6,j:p+l:p%m+::p%;#nzCxn+ øJ*Apw~Uy B#~ƙ0.oB`7 ++Q v9XJj)*.Wz F(ړJr"TTRMz*2~ui o̮koS?)!b;Y[5 I- JF(O 5i 5J*)0cq]ɚI8#i 0Hs( B!QM56_kd$n@WX,%nFeՈR ilT(6QkVۊӯJ4G:TOhxr[  Qv" C(a1twMtޣZvo߭Wdd Rw[e]|qaue#f]̖dyyWmmznW?Ii6&\]fcƿ n [fA;Z-C^?2r5jdJQ-ef1zl{leSd1O_Uqkv nJjON:\GC*S"S]6q[(.6+ qY}u!rSr&æppVͣY-oŁwѻ<D6|? r(#:t=XvB{SZ:u~4eOnapd }m9]dĊlr( p";’ҀaxK@2Hwi3臜r,<D|TFJRR#jaޱ-κ6E1X{jz.ozN3'"{$m; vtvQeM%} T|"J#![Tw "dEJ2вKKd,I2n+Rr-~ )2G=J (FCul;]` 0*cu'0,ʯ֋DN OlW(Q Ye6Ԑ\qFD"AZSZq^WHZ5 / 7*( c13̃⨍8%u͌\)-@dSPѲm.BhhKd1t^MhGy\;ayO:&yŵөL4HJ.jm'AYZ"%*ڨ !ilp D bzVרZRVca+Qe^ 2^k2LUBd"oh$yq4;U \(@XKTFI"A* GLsTu1 >C,9`^,wt$e@ 27 : ˬ޿ʆ6g-_jz# )(HeAn`85sv''xq}UؼMzw0 ͼ[S̻A `z OLi*d ıT&h STt@mNk΃Lr *uQ\,׿#w(#w|(|֧ȷ,:ĐCoFmV|#0G7USKro~bɖ(?Gxllü!R$*WO{[8qN1mGgdJ1h wI8N4}4*>,d@)Rߴl&3jZ̙?.N v"%<{Q oHp(YEܹ8IЀ~ԛޤ0wI"VpӤ`6^{ΤpH6b 7)1Y?0) 0B[>Oֱ#27۫2z?E/`}0N#HϏc` g 9hc D6Ʀa (`ZPšRb5T+Ht3xi 'a 3#Qذ4éRTPjYՒR((+%V$md-S\V'`éF&ҥLk}I.Y2>Ae:0YdIXv/#Rn}:L,t h|7_\ !HkmIL іXC-#ҿS'2_5r}Kڛ[i[|bӌu[e7jq W,}nӾ卾y(ճW?FgzƸcs#A@Z!@<b2n56/:|Y߇_ UJ+s09`: r#fs| q&N˔p ee+L2TyM 2*U(?pupڈ&̌}e~i]{O:,y[ZCSppDrO>$TLQ''((]ThX7T7EǠ얻pxmJ7Kٴ#5HoߚqkH8Gˬ:o}j!jj=;dEbV`렫X2O}b@8x^\Bbyf/ngO5{qy¡f/>ad'b 1Ӏ5NO23Dc}ōVJ\ Ҳ fWCXcwxB9<1(ea Eef&VbH''30u:#|+*\?GDT@Ϧ/hIsS9%}3w qԎK:AA:}̤kY,D 0[f:_^_UzcQtc$:|?.ܻsνGv< =;S3eV#8@ f:r' 2A"}O@E=X~o}z}zjQ9'AS[nOYʜm/>مOFv>eS*}{ rdcI'==@$ _'p'qpPws" 艃c郿 2+gE7:!mس;L .儮!ps(h{K:ɤq:G]KJXɑPu]!%ǥnd)QZ %$% 5ĐHA 4S >x( 6|J2Q P7%*uE0+k 74<7Uŀ,5mVVT6HS3Klz/mjlCe~,ziSw3zVߵP3zǛ/њ,H0`{^?3n~+qd~Ju37cbE>s?^͚rOwW޽<3 /?\Mn03CLB=|{vA~nz$~t~=L}BJ5-s(?.δe1my1+zofl|n~:^5ےko"Djv_6cse>_ z(ӨMcRifVD+;h۳~w\-/t^2kWkU/~FÍOay4i}!)aLP]±% F}Tɺ- 5dK"yBOkEةQ{9p9C6VfaϝnlٱǀJɹm w! 3SgSw0zg&mʟyS['N{ d3d3)NnXw~7Iig#Mr D2s\LSN|GĖaAdv  <_I6й$N;:KOF8eVzk76(xJqSⱚq;I˥B\ a(z ghvА[E"QM&Q@Xxy;d}?) 0j8SIQA8n?HJ Fыf=DZHB ߎE5îW0w!Qʧwp;cD E#%$NibQmǸ!%5.p̈}R#rP:^*?R)E9K^iwQfn| h,UoܥO@cG3[-ɖD.|CG% QT! xšdajAIX 5-("Dpƴ8Ѽ#rW8~Ka'AI=y'qSu'"(b(&4O?MKI4q@ K)s4'$B0ݳü@"ʝ@Gp)g0A 3# '(+knŗ; Ua&VR), K"yIJ&~OR6HI2ݧOO/2{*c&FY;A4hi3E2*x:yHre`'S=u`$17q0EnJa(UD$ɳ2#C!͍bNɍy~#fNd#vp '_Q; ʰ:>sIi/yKF9Ci4_jOF`mlfm1fjxQ.ly=IzI٢Ge_ [vUyt7?%Mi3J /]AqSfoy5"  O 8fpI9΅8-To4;Sɨg'9ś_چ:AZ641շ}, g{}W]µo^F rЋ~Э=k7_c&|^DN1EO"vWj8þX}Ϲb7Мџ+_`,u7>rs&D ڡ:J}=rm-]T8uB<َ;n3Pc_.Ǎc[N(Qv 6֦OZGϕo'ı7Oޒ0^Ut<6{" (J*3_31p}A9$ A/{G'坌7z*ɖp]a)-KME)X>YzS~opcs13vXT2QD,Ao;$LM8 ?L0!A'pz`d9̜AJo줖ĵfc&!Z[WvKni_-zȯX"!S4$⚥i3B%$ 9ƘX|p?G.sM>,Cj惪>.b]XނI& tٜMuʓZ9S5'OES(?\+["seKd%2yQ 摒Nedb 8:DT ? S#DSx$B΢\uJ\uWrDy^/B(:wx}Plre+\8!M5T'aHCJ0H pࡑd|;Z\`baF&XqR$OhR8d (k TCP`y*BHs6l. lgiBS2R(RqD9D%a*:)#J#B0HclU=FoP66r.u.Ő$6<"PI4Zvmn ]2H>{0(t%|0"⻹>ytӅ73x WB=_,p;C1Oq7@LQDlC2I"pspiar0HiN!H0{,R* mec\I 2 X*JJ0evqH,D*qIB0T0jU6Mף(0YB- /Rh;AR }FFjziߠJb$b3*ȝfRo(62d&$ 0!PUj] McM"I_Y(_Tg]#kR U J%^8ff) XipXy`p٢QTZ\Z YLcɔPZGX14o٭5pXr('3FhL3>B.$G  v}FAsгQUZP+ ?acw5~|{ 8JUcn \ ֹ㷣F/(ݿ$ҹl>7A>WY/*tJ v}C"j/ӖodnwfqºM5AnT~Cܭ-+i卝B%-jr %{|viҘ'T$"$L4'ZAdRKc8AL$61 ȄƠYʌ5 ;^ci\p0Dq˔?)CH~عuμA#!֩Q@cmGY/w`Hd8(6i,4SP')I 1J9Tȏ ՊW&՝'T˝ٯ3_dSU+3_F4tSiM? 34̬<~7V }s#u4`M|7]d#ejRWOo.zͧvq;"E:wmvˎ5(vLe ZtLrZ RTWR 3TkXZgoorx޲CED ?4ՔS?4:j|F^x( hmIn`/m3%v<&lfng|l;ه>}b9?:\ړnB~9 ,[*Ma[Ok#+T8`J!*„!"LL2qZUp.孭0~8rT3OΣw 8*6Im8!R:L>ȓؓZho3PmֱgRED-kƶJ9@ =*EO7\-Jpm{)̺V7Wl5G`ޘ!dU|n`zj bYN]/7o0ܞxrWdx6~"ƘxvaDȌǭGx1Cz+_˄]dd/vAQ5"8gF— -S!34,Gn7v]&N? lf|6,ch?g!* bv6v{{ oW,fw:0\jvsTmKUh,EQ.@!6+ےJ&8ŃĎkRE(ϟeiv{Tx s^'Ů f` $28DKۍ !v0KtB[l[pͮKN²9oL`WѢ{y-l;oK,ٛJ4 U@hEv͸ݹfQΛ7dB'9,S6<$%!) _ E /'6X=uuu]]AvsƦ/H^žX{16a҈Vw_w&krjHCԠ&-橰X':9&fxlʌ(4fTZg&Vf~C^Nν:| 31+]?AyiM~6[l7H `wloɞ6Mi$8S­J:Y )2:,!HhJ!fLY(-lF>H(fUtQ*HHՊڋikU2\PֈɊs:>'q'6{h%VSY2ɪB=V&ѕӽ30-U&Gs=Uv,3HֳnlYǬp,v$skLTgokL1p[o8qqV$yc4l\|p(K8GuY(8ZFs)Ӱ(3J _(5_CQ# 0oJ>y`(`(n V#9%oJv#ߐ|Cc$gLGYÏ{I Sr:j(U]v_},Rz_!#1`#YUA-t o2;G1jԏ(ʈHdؠ\Md ]%I`\) lHNbNɁtW -LB4ia֠w:%%=OrZW) 6wfx䠨kq5L׊q񸨇G[k @Cj鏃gb< q"5y5PS@1Ġ~ %ĠxwHDǥ6[6JuUr-Lù ,ށ&L  -~]tJ!sUkf:% tIBP۪C2H[yr 99#WәɊJŸ(HiȻWO3F?̮NWR3]ѴXrkWEcuc$2鰷λ|wʟ7TEJ꣦@U*-W [E#5vXC5T XGF9SyltļGvmᗽ3*S٠:{dwEf96iJ/ttxPFU]!%cv@1PoakJw A_޸Y7.? F[y*7*ejT`Lj ?uUUTjjkgn3*cXJJkV)d-(cGvywY(cbn(}{6Y*L?AnKkG,ʨv k,qV\P)\S Gr0c4kmQԳ30%?[-jJI:S|Q0O1"1Y'T8ZPI"iMf?EFQzM)fko(xlx(22wPHXP7;9rd2안NX/.i);g:*drP :eFcd(ǹ͌]L(Vh5Ԃ^F CAHgxsMF2؅ZCy3 nArV-fPf!L85 ],.'?876Lo{ؑKVf7%Uc]wuBLﮊ< t73?r5kCZL%׹V$Symx~qgm~83bY7mBz Q ZʮJ\׎ELiM|[SĈNwTn)Bղ[cDc['F2jj7^AKM 9TbfGqT Bf9rһڸ; շvn J)쎖xBE'G%!EZY2)#1eV˼ %wY!LVoDx}!epan "id^Y!be2Bf֛cop@Sp8-Q 7c@ 9/KAaRwegzc9"SVpafB 51݂gojvgCuHIEH.o}H%Dn6fj@oi+&d*3"]IvyjQѓ=0Y^SbKy~[$Ő̖X$`ãQwOu!4/sN\ɋp5}.Ep5`~0& l:Қk";03]W+cEcSOx"ީw@K1׿:{ŖME, %scw$3@i#.z Bb!ߨg/NEYB:'!xw1o҇qNYL$h _y[5ww0hvMԁ٥D, A? ih u8愖YyCT5KA٘=P9 "NjE>i~qz+k{0 L/-E? Xp?U|3ʹHvT= /_%wj-_=_Eg"~,moi!#[ +Y6`_!ƟTþcIuU,C ѝuW6m=}hwݯbBC $ngL2 1xV| @'1[{s@ȳxӞ24I/V23'CV_zXaKQ)4XsⰂ:C| "H 0rg$u3 ɐ*y򰍧/M=OWMv̡q9ՠgpD0II* $mw$gOe`)g+F5xa1n {ˎA&ܥKjyz%됲2ɣR~Ԧ/`0D`ΕP,`#;VͲM(éA?;DJǛOgCgDihboA 2D2vyhB;֖2?ïy:QR ĉAS i A3TІRą2 qpdh!ظ_n[oVB237[:~ x_9|,uyTdjwgkWw=_+sW>OO:pxd?-e܉ga-jJ` 96\bM-({kZPMIX~9` #7[ECטd`AA_:8c;|Y': h9FZКaD('?ݒߑu-.GZ Њy-"BUb?rt-_rf=h<ӱ2v!=XOyZRr``M!!o\Ddl |M?=Eo<|WFw~Wrgt/ gNՓFPBUZGC\/dgICNa> noMxTFzcʐ#$H<L;g2ܤS]~m{7?~].ݑGXM7O{AGg\gݝ~gOϗ׷սg]?QlXvx9[oc9k2Kn~ Sv3 1PeV-}!"$yHsE&Uz=838gT8v,Sr0dUh9'S@Btpwgs7^]3%K0ގ[5O0{fIU冈GV8!M9=CrJZ98z˭{<<ĞTC _bN3AC^NGO[esdE/4%+q-S@p, 5tl"N7;>+8&.P rB"e+ vfEK3Ml.뇐,Lx?Z/kP (Z 14|K);Jn2C1*E nD6Py}_:PaaBV rSik){E.6qa-D62i Rra %#\3.٭Zc hP=P$bA&U 5)*Y(_k2h{ 9/XVCI< :D9/H91!؉3|/Bj.A)O3MGmo 1.sƯtr3*x 8zrtbHtBjMxC!j k`@`ܹs1gPRh~#;V sP2I>h)X4XHj3,/[XԂaJ4 |SpT줧7w~ C{)C.Z ^U!!S($ j,r @NgSq-Kx*8t5<#;HcQDIk˃`'C*zDgqHy|@ҪjSmm`Kttgނ Ir U&n5VA +?b d6PSZs.9HRc  K4@D!XF@DP8QXFAu>T3D\/3ɔA@_o\p?~@KْjZtd"߉xgxc"@|e:ؐD@ {H\  @ܚ=Ual#C ^=tb30*s''ǥ/Y0u~=֡?/Y_? "h???XNjKB9Ill\2 0^٘Ly$L~TC@84C-aF)H/1R 0˽ZV.ɼin/}-?cl=p%kSQI*`EV `|œZ\̯z /\>X<"(B3rMu,r4;R"k/> I/L0̡Ň|! h%UY|XCB zhXM*i@u{֍D Cbnoċh!__G7ՈS{ї#%MJa4 ,EU\]5hݪא\[2Zݑѧ{UM{Cn.cG`bF -$9#'L1S~3~w;C^bbVϽw'HC.4s\2aoqrC+btЕi(!#g67?\\hr9uUL F b95z NQ Dl9iժ~Z.b1ZMj[ܚlcc^VxAaE7pg~?Wz [d 8ozSmE| 4dçV<ռZg,ԥZ>LAJyf S sLrYQx7{:#MF\8Ivೲ?=8fOTpۘ$c˺ٷGIX([4XqҤ'գ"NKI,Xhjqظ]^v(@sn)f1뻗[<<ĞrY.,>dt)"BBACUO ڳ3^:].#*/X ܢށ!HS?G Qϡdˁ䆣| *phh&iC_̱|' , rJ_E"Bh]NZY/qAPr.˹L`jtqfvQen Du4rPNx}e:%ɉ#xM4'$pOZNS-A13==pxS)oO蜲=ʕ&ҝamSA6h+m5I1jl#xC( v<hH)hwmq:yٗDQ4 ؗBױODZ9@ܾnTUq0I2UGm5f?*|Tz[$j⑱mTl¹SuF:2^6#8:$V<:UnFW]GFsl+9*j==ʃd[4ne3## 0iye~k<" ȥomCN8J3çZKi}PQ,hE0*Yʨ%] hMYkZ%:W ƬjܟXn,eхbQEZE,eۀ֩M!V)6A`iz=r, ZO K oiV_(|4=BDYEVt[ ߆UK<:*Rb=$`ԒPX+^6{A:qhQᘡ ` !C8{ت `чNZAbP#c0fȥT1CB~smPmi7(bP#:$ݚ nsHo.M25*ݚj Ö|wĢ%(|wEkCKʦ[9kV뉽M @!\$ LԁB a:MɺlV RZTXƹi*4 J hL.6s3cP cUl#`5B$SE:f]nϙT })3{Jȗ:̰yܭBha*p] ѠP=*fxf&K0񚴜4a9ܟBQI)m'HR %fpX2=ݳ*fxf|cl/_1.U%0V/9!}do/?d }l.;v8j!n?n9c(QGL%b"2hi$oh:g5tFɕ[ (nсB(.JY*:OEH d41Q٣)) )mRLfiuDjuԧ +~} "koPL.:@VHV-]]_QI9QyP ϲc#KJAPe ĤmT6>5]TzSz9ȊcE#4P%9VEŇI DK:Mұh|*v2ؗ1*]_Pn7[k .n9.Ʌtbܛ(q%;_UHoM}] 9~av{jg7Hzn˚۲ʚ.॒ L8Qqn(;h<{f=+>[ϛѫseU9V{k2-F)fj{%߸5s loːX stM|-8pu]EѰq`+bzچnFк3[Z5ֆn 5}&t#g;kA@^sCGEU+lguK6DX9Ue L R!I,ZeR.Z$ 6tINLnUA*07[țGohP W1-Qc*Iاeg j@OeQ8lq^xW@K1\YrѴ(}+Nk\X)A7_VMBCO+5Ss82U;lV '[а!L$J `?׊t2Ԃ+DbP[1ƂPo1OIfۣJ >""SoG B/]XPwRC ֩%kfx#&>%7S|A ]6.@~v+׬,Kx>O_5Yt1t?..rF-=݂ #Hz>xٹLEBB]YXn.ͻ%R`Ekr(٣Ğu1Bɖh.A LHiOq UOo%j_Hz6~JʖY9; y(G]XQK%0& .9#]P".B!Y I =.#`}EdLf Β֥$+EZa jEN4UP3i5,IJlےU᦮3es}c c=`YeFXɄ)hVl 6s W NjޓAؾ}޻ѹk]3:cfses5Z ^u}/h:9&DIWӚpī^c?+vޢ5)Vsb܌kU5Ϊ\57X_k+`vfYLkZPvPSSqA捚pg(e8z["Sdq ؗ1,+?DUi8M Cy XHTOJ{ot9QFiDL|YVMCWEQEQD\wls?U\Vp6AJAb.e&{OޓEȔJDKhԈ#;B _Ò0 U{j)CS+5)$t@+FrjY+搐\DdjBhu=V%&lN9 1˷Bw[ho*9 xNsU[3" SZ׬UζRٷ :>[rxDO»^.N]\<ogy&|&o>oL_)RN.nA'߿z{ө6y}7d*}fОG{Ez D cW]_C.ؖvĬpQwb#v!^ rsCyB ;עx-j:d³r_#*X8rS]2 ?۵2>'v/4it\ة$ JeŃ2N%|YiiUQpUzpIb>G\lLH> ms(b$%ɂ7#5rbVQsӪy'Tጼ׺hyQ3ƦthsWAos7ٲC@GXF{>DI P)Q O9 If !')"ܰj$g5Xze !iJMh+ή +3n 8VVd#~cLAES No*琐\DTN83g>$';v CMJ %sÚve!^FΕ_"f9iE/_3C;[$-<)hMs!t)-;? &N"`9Ujd|Sњ ΐGeWD;N) `s81[0L.6EɡlqF̒w,'{VY;Ef*WzG6Dt)-dKGt{:@XP+YrD4>:MjEY\w"9/G'U}bΟAt`iGg>!GMZEGU>jE5pr.J.iJ 1HN2 7diWWbx?oWmn2q _ePM'98BMPIu÷`K|} 8jȍ;ژ*9Bt,ľʫ8=)&>gz+Դj$"]aGa3-[ qSTF,X X~3 ᥔvԍ!5sKFZݙCLsɻNQE+ړIZ3g&}ډ7V 7_"%J"rp"{n!8VIP@Y~ڲ)Dݧ=:HJDCه: _\`РNvRT(L˪5ĉb3zQ u,^QkVh@' `ԗ'S aY;q͹OVm*WmZD0-GHa 쭝s@o_bgOy%/gu3i_z~9GdWW$rܬ綖Ͼe>o3_suz}}?w =;4ק}B:O7H Wy׫>)>.|Ndljê?#BkTnѐ.1XMY *sLwmM$80;~Q` \vQWPXdj[-e cj/*3*\Dɥ/|Xnf6\(R_\p\Y|t8bE'Q˜K{ik4۟ m#jIZfw}sw y] !HlE . "qzDL:%,rg`p;_;1+qK*˘껩r;p::8sx{"uۆ3-%jW6__$!NRkZ\z;nVXnnzkFQ}F̰{WeWZ-xFX%{ QAԙ5\q.%sK+\;MT C&)KQrXc D/8tʕKǝzǴ>|_>pvN<&;֘VnJn;y;%qcU|2'ھ/ʎ?=zs+c? +&1dq#r$4}Ɛ~/ɸkwYhpNh-/޼wn!R12(#nD67`K͍mxn%w$M TK뀂Vg2F݌;ʩa]tI:0&E=UdQ\0 [,oxϞv 0ݶ3kԤꀀ##a]DvȶnΣy\Oitj(1^s_ڌ樺(׆CZgB %^J$ IjJ A;qR]Jh1!Pm&wCf9.E#M 3a9 b<0xD΢-JT-$%J~[5hm,ߢdr+*iFtժaГE&x + {3;e+I{fvSa`h@@l>J\fFU2k bw[uH%-߾BFxPv= ]ّBDeW+&Z֮cBPaɮ)Ft~ʮ-lW&'}!x!FEyZi) KoL7C{Y[UQqy_{BCL|Z82bNxc6qO;Eݲd@٢2n[էW) o]?o^<9nUml&Fx\O K'eټTgݼr)K 6O)Nn+ UK6y~Q^N{" ID[&*Cu?oȆc,QOѵ\ܚ ɷC#RteQ08Q7Z_< DaQsb=މJ1A ]TUS-!F 02 npIh BGE+a>a͔H{$$J!XȘ L3 $ );:D%v1ATAkHd㈹^;ae):c mXen/W7):iNi.JrvY,͞"A Z}Η#fjZyc°H}S燺SEb(MRL%UawÃ4%/T4 F*>= tO)\Ã/(J}ysfZpI1<+{^Qx9.aYB-[P :ln{Omh+*t\Y6\FE.r+}TtLSͨ`qY>;+\2P2!TGf@Z1"aGog|JsoZ p(Ƀemʲجvg` gq܇K u1;<\{1mPBV8y3~!FyT|)k )ңǓryO{t2{7#[TV޲`u}1y+jϿixu[߆'Ydy^%_\3x2.-s޿|k0y˳hX} W|0`zq>z ΋8=PyJC6</'_pOg}W(`xOSgx|˚pK J^ϷV乞RXFZZׄe9{׏@+zrҴ y]b.*5W-K) ?\{gR?Pb26WLVvw24k)uB{!N]r؁6w!ۍYFU8C{˵Sl9icL?GozeGVJ =88OL3*TJ*:f$״;Eer4RC܁=>("- J0y w) 'Q\53K 6Z8$QZ6ˬFzJs%ɰhEl;AkϝAJm<&7ũĂN8T$]gڈ+V{B6TDY.aE0h''WD+ˆn$~rE0Fi 0 _"$f<0EqA W1͇OOr!D1 ^z-J-#h"&qsjM0ſͮjVSAm2vbh'xOrznPQGp )mWAԖ嚰CȖ2,K[2ϼBx Vi EvI'QهJ#P i-4|Ҁ@IIb:׫{R=K>Xt}̀|w6|sҚY\@Fńbri.Pޚv#9Z%Vtn;Xf6`{fDj}5⇩M6|6{̘Ȑ"](gRB;)eY&{Cz {hB:ܕS{6iқ2k#T嵷\ʦ`.bCP#X**N%x?{Ƒ K/d7J)˛JMKRRBQx E )!)9ardAh4|$rgcJ)s'^kpogSe`*'C+Q7"'?"c_d :#f H$j dT#@W>n5W&O1b佭8c*zYw8A|(w))gZ}.'g0 VGd_H,ی^R 4|8dld/%9B]bW' ؠAY1<(^IH|M+Qj53C£oȦyAn2#JYeGnv==D,oZ' (P#+RcTD2$U1L`rQC8MyU4 =A&Ȓ1Ɂ65J;4'D(ȹ042+4 T+jX* Ϲj8mH"fOIyAQ-}d>&d"5 kОۣ7e$C&Di+5ӄuh b륨2xGdHRqWD`^pD 1x!y݃j<4A&9ȑ4p>P+83IH4|Ն@*3̢D IFM2USCAu 4w%"qFW-xu²CS(̈́\;Gc{ a#΋p #o;S-8u_>eOOlpFuE4ꭚ{/U_7sKjo\TL&k3`*no>fuqޛޖޥsx8wznF6|UMjdL?uw{p\xLGr2j͡ho>ldeD8Ɉ|oԇnTx~ B ǣsJD[A1)* T'L3b*']JtDPHcҹd#ߣ'4 ):cpal=e^qgoQ&$xlн7Qe M(5+m46abpcn p onHʵ|4CF;&Z9!:Bz5 L:XA\FW2:ղj9" =o(5&RD@0j.$Kz':9m\{aT2:XZ-@I7uΈ)EWMr{e`HcȱnzoL\U=boJTJy63WAJWmů+y$RZ AA2@7wd2L_x6WVb3=W8e2/C"ū7|`oooP}Mǔ7 JUP/+MhhFMsjO>LݧqR>!o!1jIo./9iB=~+"ͦWɾÏ}rzbw0L=n>_iJM_ToM)zvg~>f`1' 醌ȅ›W_݁F"H,|"[Ƃl>dgr6Gimx*!ɍˤ-ϸzľFQPWEAD($pBꓹ_$q}~(E?!7۶US&-$r\dA 4n D{ton6eVLsw,e" %y"V86\"ֿ-V(]v ɴt=,IW4)*<.:IcGեA}}x3"7Ys) 5"QyTD BU*. }(j=5PJ*.wen.ŵp-U!A7nlT\yUe Z|%{yqIc@]b~_zC:l%͏/Ni+ uHɩ;( ǽ#8`A6P10S\&o S(O5 v[QP\1FETFof["(d!i'wh=B-ڍ%KơAoB4y-qɁ٫=/u?2eW EfNj \ :34?#)TE#sbPݹbYQTǶf).e {mx/n(?ܜPե7_W}^Aw*T^,:}N3tNp_p{ 3;ҟ QиgK]u(KFI.:i 䘋^pĭ\Oᳪ6:"$&܂jc4胥|NDR\'+Rp-)퍹} vk&4w~P7hdN맙speo [>U=˶yKc`ga4qGC6HbzTJ)c,xFDO½%ߋCu}d{gt aGZ+TbM.Tt9ŗ8ŹJ\iEs-g&J6S˟!yNQA/VEi(8]WXQ w߲AnХg7&4{ _[T/}FMyY0K ΟeW]u!@X8' W`6#\;Ґ?))>غ5dкA}FRЃYυm'W^:ŞȺiָCzB֗%mS"4`֭DK.4OuH.v?LGjS+'݁A$ Pg?4(S FmS!PLc?Pv2 Մ}ԦK}}l'ݧ we3R~8tp2Q"42 .N8+yfF@7){P"lH7" {:0y{?e{Jeeu4+[Ȳv9#JOC$F1|ϡ`$n}cˬY//^^6r~6aA8KLt g٠nהc+ }t6X)kQ8~xwe6e_ʼ}]bJp5[e.q#j@Gѩz˨{hV w-rQS˛;w~rN'cwxi^xE#gtDq2"UX sDy F2zoT~DZjQofk+d!C~&Y][o;r+/"yI"$pBͶZ#Sh.2{2=aS7Vɪ^i9R,[Y>96~MPn;[m>uڦ< وI|@g}9z-6}v Zn h-2ͅ i< 2s5PL\g@?"iq,N3wF`z1:|L8jI,Qo3;쁝f7; SI__6$B@gE%IÙ䐸s:9v n1DE[㕨K@F\OOEPE~&N.AJM6Dv:prʋR7/S9SZ1txy2 MJN cWԤE)ǣy,MzSHi o x ӣkq3 djQ‰ ZO[ڌi-_V2}9zP5j^8pRڻ}Pu?zJFPN_A3c-Et ;^ys\frr(qF5w;ՖߵhmѻS3yGrJfILԿ Jӻ~GP򐭡A?}Xȑh'|$Yn^i,&yݾȒeS#,銐Q<v)rw3Hz ޭ JLwnEѼ[wBև2 !N_+y;ULjr@I"K%t/k5У3F8^*'w4hks(Ǘ-^:Q(ViOֱƗg-]t9',R4%ش]gR6h? h 6m#'(r`ӺM6s ieiTSã֢-x<ش=K1lӣkq3=ШFe}pnRDLFzRK [^5a̠ZװbNWpЭְ4NףR#PPSUˑ6z/0H]Ы aGV.l-ֹzn-./aKk.ҹ  J5JRăh9×قEf.8Owr =[GQWn˟^aSn8ߛe^9|w> X2.>p8;xpr%S8G Eu!޽{vP?~ gqyB){!~}CY6g<. FQd/ws)9g4c 2kWD竇G 1T._82%Bp$&N,jU"/ޑCU*W/Uz,#gp½jW=t@u7o r-tkEUMPgvY~5XjsNjѵ[hR=g<-tFd}\.{Z nvɃ GM s[]~;կ 2Ȱ624j:K|dҔV;֣=YyzFCYEݩ}*N-!o Jpϻ~Gh=^@ƍ=*BD;N?R{Z=z2(1ƻ1bty(n}Xȑh¥kS8׽-!ѻuAt6mGΌݺ9z>,MtŦ +)2 粣jKް)3fFNRgL<+Ɇs}b:ooǁp>'r%Zk+8w|z^Gb=O䌉[MaMdEOaX$7[{ KCP{1y6JaH/ G܇*/_rMȾ- bn4K-V=X1wK]#~h:Zo[d^|32gx<^i=>RlM)e#8٩:Jupech *,;N4%)e''zM=W_97g>t ;dzw+%֗mmlmy{bNYyn)~u+bY>ۅ6g/U,\rik9nw/!7r}KcK=ߊ K,$S^nZ'~[MK^;w"I~{"QcvHsH hXZ2#+:-ZmEugaS}J5_a05?O}{P]]1:HOlҽ@B_ 5~;?s>NҼP@F(* |aD+F󌽥> 1]y4N;౹~.,];왌U)DѸJF^Nbޕ/V*?mUwWz}BK__)gwA<~gȾ>u9ؿBI+}Tb-[ge-M9a <H0.Jyc֋rUҼOp1W9kI]HW@^̽ח7 OL G2ttF!4!-|PI  DG*/̻9كy4NGq )ɉow !/sQ3+RĤ+9(|ݚ]5/st7gMEf+ +H˪SNbo4!.UWqւ2 D"HE:4:OY;'YR%%9n ^tj"z182_NZ3 +}5{:p+uVflfW>N.Zռm z*urd:^E?Ev-T)@jĩPU%$cl *-[wz1#ɿޥ gIԫ@ җ<,Eoc/n@ d~5La  h⍑1U<﹞{+!&\wn3 q;[FhwrIIb^q0\ogtSXIkr%$W[ٿvwh_oyiOhT-$Z3љ[OFcѭZZ<ێ_oE&O0Zy͟ĬL#JӊaUdӛrH0 ߧzw$*)*OAVKHQ֪J.MZtRn]v4E= nOnedpw$r*iƐݞtK&k%G9:E >7_?FjB-Dh+zr q f:]̾dmXJy9- Qհ[82fΩ'ZAOTE!0$* U:=R{t:J'[ R:ǡg+U uoUlA^i{zV2էT[RJJ-ReuVէT[4JJ 8sӉ/S}JxVj|=4|eVH{KJAY)ZX4]LVJ/d4::xmk:5%mӪJW9SG1@M`1KemmE`0;V~?y"dpW*$!]e$qsәu!zPOVum<ņ<` cUn1%OW~1Rkqmiiw|݂n"h|J`&e_ b|ve6 X+GNU>GO׻ qfQM#O&YۛcI 92H 4;$Ҝċ 2%x7cz7_RGqs͜=Iw_E}qRK ~["W ?C㿽$zT3oz_1Vvy )H@ >tлi'>eInX!93L= c<}S'Y~xyG!T\z~Ï:^?}f Sn} *s=Y'ŠպDskͰd->1ŧɺM*z~Oeu9?F^u4m\2*RL=D-Kܸ睲z5{-_LK k~G8̉ NK`"~Pj3dkۏgVe %(UG;R3ՉC19eooō]'Q`h6yǺB -ꄾuksĚn縞֭qnqyr gw*6?NW:Y<ϟ[ٰgw(JK%עon ]wY,w 2ԣmA!t>>SiG,XH8MbLj\YJK)H7XJK3Bez,e,PP`:[lsq@Ѷ9nyKjT=?ڸ'a>9nlseQ?f B%]~0h64yԓyӒ*Zu(D҃8Nɀ5 "DX4KDh;)ap/;QY j1LL@"2RA6"5+N2ÔT]hTe!zXaJ]oL-[M#U *uU7h9:FM1R@ƫMZlxFWW+XVJ:C wD*Y_ovxr.k`tt^(ʃj :PE6A r{s#ѩ֝-a"ĩRɠ0tw=Z]nj˿r55Y+4b x,ft ENyK~1VNσ(&cl]!Idӧwf3JkFlhp) UԆ\:& C,`Յ6Z|شPLJEH!2FlB&B!94$I)5@=%Cy%kp*pCorq!SP')4IIb{J hH1JbJ& h`aT1kzXs&RP'Ta( mg|1dC,RD%x9dȌH֋\}>" Oat=j>?H˾ٷ~[2͗YdÊcL0{bVmƣC4 xy0 I+~!ظ(TDdbe;VW) @pn&'=<+zoA:[ԀΉ1lڨa!d-&qyjˉo\=6碬QE/<\r c$=r' ϽAnsUݐ:DkNj~%W-`urEyKǶb-g֚]X%f+1~ueKj"%ysE[yƵX,FEiHmdr.Q<^0ix(3nX nhn8.<dža8܍;,% %[;_t@S?,t7 4=@j@حZݐ,#'<~ +҃㉪KmLxfp4be)RMhGjI:Q9(SiTJ7=`hVIv'N9[(> }G֞w`u*+LVi`hǹk݀NX@ R֭U`Y-_VC8E8=9ŇXI&Xn5v u#[ @DRvVpNo8Xi 3Cq `)t@2$f)U~,X%*K_z-5 rYpeٹp ci&ə'R♅aYr*;ץmKm-=9︜lvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004207447715136707101017715 0ustar rootrootJan 29 16:09:52 crc systemd[1]: Starting Kubernetes Kubelet... Jan 29 16:09:52 crc restorecon[4699]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:52 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 16:09:53 crc restorecon[4699]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 16:09:53 crc restorecon[4699]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Jan 29 16:09:53 crc kubenswrapper[4714]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 29 16:09:53 crc kubenswrapper[4714]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 29 16:09:53 crc kubenswrapper[4714]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 29 16:09:53 crc kubenswrapper[4714]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 29 16:09:53 crc kubenswrapper[4714]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 29 16:09:53 crc kubenswrapper[4714]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.941267 4714 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949558 4714 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949597 4714 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949607 4714 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949616 4714 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949625 4714 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949634 4714 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949642 4714 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949652 4714 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949661 4714 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949672 4714 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949680 4714 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949687 4714 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949696 4714 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949706 4714 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949717 4714 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949725 4714 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949734 4714 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949743 4714 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949751 4714 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949759 4714 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949768 4714 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949775 4714 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949783 4714 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949790 4714 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949798 4714 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949806 4714 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949814 4714 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949831 4714 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949839 4714 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949847 4714 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949855 4714 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949863 4714 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949871 4714 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949878 4714 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949886 4714 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949893 4714 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949901 4714 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949909 4714 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949917 4714 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949925 4714 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949965 4714 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949974 4714 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949981 4714 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949989 4714 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.949996 4714 feature_gate.go:330] unrecognized feature gate: Example Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950004 4714 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950012 4714 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950019 4714 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950027 4714 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950034 4714 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950042 4714 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950050 4714 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950057 4714 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950069 4714 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950078 4714 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950087 4714 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950099 4714 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950108 4714 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950118 4714 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950130 4714 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950142 4714 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950151 4714 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950160 4714 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950169 4714 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950177 4714 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950185 4714 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950193 4714 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950201 4714 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950209 4714 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950217 4714 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.950225 4714 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950360 4714 flags.go:64] FLAG: --address="0.0.0.0" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950377 4714 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950391 4714 flags.go:64] FLAG: --anonymous-auth="true" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950408 4714 flags.go:64] FLAG: --application-metrics-count-limit="100" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950421 4714 flags.go:64] FLAG: --authentication-token-webhook="false" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950463 4714 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950475 4714 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950485 4714 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950495 4714 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950504 4714 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950513 4714 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950525 4714 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950535 4714 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950544 4714 flags.go:64] FLAG: --cgroup-root="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950552 4714 flags.go:64] FLAG: --cgroups-per-qos="true" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950561 4714 flags.go:64] FLAG: --client-ca-file="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950570 4714 flags.go:64] FLAG: --cloud-config="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950579 4714 flags.go:64] FLAG: --cloud-provider="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950587 4714 flags.go:64] FLAG: --cluster-dns="[]" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950599 4714 flags.go:64] FLAG: --cluster-domain="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950608 4714 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950617 4714 flags.go:64] FLAG: --config-dir="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950626 4714 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950636 4714 flags.go:64] FLAG: --container-log-max-files="5" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950648 4714 flags.go:64] FLAG: --container-log-max-size="10Mi" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950657 4714 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950666 4714 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950677 4714 flags.go:64] FLAG: --containerd-namespace="k8s.io" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950686 4714 flags.go:64] FLAG: --contention-profiling="false" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950695 4714 flags.go:64] FLAG: --cpu-cfs-quota="true" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950703 4714 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950713 4714 flags.go:64] FLAG: --cpu-manager-policy="none" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950722 4714 flags.go:64] FLAG: --cpu-manager-policy-options="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950733 4714 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950742 4714 flags.go:64] FLAG: --enable-controller-attach-detach="true" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950751 4714 flags.go:64] FLAG: --enable-debugging-handlers="true" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950760 4714 flags.go:64] FLAG: --enable-load-reader="false" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950769 4714 flags.go:64] FLAG: --enable-server="true" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950778 4714 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950790 4714 flags.go:64] FLAG: --event-burst="100" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950800 4714 flags.go:64] FLAG: --event-qps="50" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950809 4714 flags.go:64] FLAG: --event-storage-age-limit="default=0" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950818 4714 flags.go:64] FLAG: --event-storage-event-limit="default=0" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950827 4714 flags.go:64] FLAG: --eviction-hard="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950837 4714 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950846 4714 flags.go:64] FLAG: --eviction-minimum-reclaim="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950855 4714 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950865 4714 flags.go:64] FLAG: --eviction-soft="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950875 4714 flags.go:64] FLAG: --eviction-soft-grace-period="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950883 4714 flags.go:64] FLAG: --exit-on-lock-contention="false" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950892 4714 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950901 4714 flags.go:64] FLAG: --experimental-mounter-path="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950910 4714 flags.go:64] FLAG: --fail-cgroupv1="false" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950919 4714 flags.go:64] FLAG: --fail-swap-on="true" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950928 4714 flags.go:64] FLAG: --feature-gates="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950967 4714 flags.go:64] FLAG: --file-check-frequency="20s" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950976 4714 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950986 4714 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.950995 4714 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951005 4714 flags.go:64] FLAG: --healthz-port="10248" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951014 4714 flags.go:64] FLAG: --help="false" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951022 4714 flags.go:64] FLAG: --hostname-override="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951031 4714 flags.go:64] FLAG: --housekeeping-interval="10s" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951040 4714 flags.go:64] FLAG: --http-check-frequency="20s" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951049 4714 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951058 4714 flags.go:64] FLAG: --image-credential-provider-config="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951066 4714 flags.go:64] FLAG: --image-gc-high-threshold="85" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951075 4714 flags.go:64] FLAG: --image-gc-low-threshold="80" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951084 4714 flags.go:64] FLAG: --image-service-endpoint="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951093 4714 flags.go:64] FLAG: --kernel-memcg-notification="false" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951102 4714 flags.go:64] FLAG: --kube-api-burst="100" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951110 4714 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951120 4714 flags.go:64] FLAG: --kube-api-qps="50" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951128 4714 flags.go:64] FLAG: --kube-reserved="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951137 4714 flags.go:64] FLAG: --kube-reserved-cgroup="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951147 4714 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951156 4714 flags.go:64] FLAG: --kubelet-cgroups="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951165 4714 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951174 4714 flags.go:64] FLAG: --lock-file="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951182 4714 flags.go:64] FLAG: --log-cadvisor-usage="false" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951193 4714 flags.go:64] FLAG: --log-flush-frequency="5s" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951202 4714 flags.go:64] FLAG: --log-json-info-buffer-size="0" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951215 4714 flags.go:64] FLAG: --log-json-split-stream="false" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951227 4714 flags.go:64] FLAG: --log-text-info-buffer-size="0" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951235 4714 flags.go:64] FLAG: --log-text-split-stream="false" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951245 4714 flags.go:64] FLAG: --logging-format="text" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951253 4714 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951263 4714 flags.go:64] FLAG: --make-iptables-util-chains="true" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951272 4714 flags.go:64] FLAG: --manifest-url="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951281 4714 flags.go:64] FLAG: --manifest-url-header="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951293 4714 flags.go:64] FLAG: --max-housekeeping-interval="15s" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951303 4714 flags.go:64] FLAG: --max-open-files="1000000" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951314 4714 flags.go:64] FLAG: --max-pods="110" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951323 4714 flags.go:64] FLAG: --maximum-dead-containers="-1" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951332 4714 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951342 4714 flags.go:64] FLAG: --memory-manager-policy="None" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951350 4714 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951359 4714 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951368 4714 flags.go:64] FLAG: --node-ip="192.168.126.11" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951377 4714 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951399 4714 flags.go:64] FLAG: --node-status-max-images="50" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951409 4714 flags.go:64] FLAG: --node-status-update-frequency="10s" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951418 4714 flags.go:64] FLAG: --oom-score-adj="-999" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951427 4714 flags.go:64] FLAG: --pod-cidr="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951436 4714 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951451 4714 flags.go:64] FLAG: --pod-manifest-path="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951460 4714 flags.go:64] FLAG: --pod-max-pids="-1" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951470 4714 flags.go:64] FLAG: --pods-per-core="0" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951479 4714 flags.go:64] FLAG: --port="10250" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951488 4714 flags.go:64] FLAG: --protect-kernel-defaults="false" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951498 4714 flags.go:64] FLAG: --provider-id="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951506 4714 flags.go:64] FLAG: --qos-reserved="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951515 4714 flags.go:64] FLAG: --read-only-port="10255" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951524 4714 flags.go:64] FLAG: --register-node="true" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951533 4714 flags.go:64] FLAG: --register-schedulable="true" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951542 4714 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951557 4714 flags.go:64] FLAG: --registry-burst="10" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951567 4714 flags.go:64] FLAG: --registry-qps="5" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951576 4714 flags.go:64] FLAG: --reserved-cpus="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951586 4714 flags.go:64] FLAG: --reserved-memory="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951596 4714 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951606 4714 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951614 4714 flags.go:64] FLAG: --rotate-certificates="false" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951623 4714 flags.go:64] FLAG: --rotate-server-certificates="false" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951632 4714 flags.go:64] FLAG: --runonce="false" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951641 4714 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951651 4714 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951660 4714 flags.go:64] FLAG: --seccomp-default="false" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951669 4714 flags.go:64] FLAG: --serialize-image-pulls="true" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951678 4714 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951687 4714 flags.go:64] FLAG: --storage-driver-db="cadvisor" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951697 4714 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951706 4714 flags.go:64] FLAG: --storage-driver-password="root" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951715 4714 flags.go:64] FLAG: --storage-driver-secure="false" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951723 4714 flags.go:64] FLAG: --storage-driver-table="stats" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951732 4714 flags.go:64] FLAG: --storage-driver-user="root" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951741 4714 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951750 4714 flags.go:64] FLAG: --sync-frequency="1m0s" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951759 4714 flags.go:64] FLAG: --system-cgroups="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951768 4714 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951783 4714 flags.go:64] FLAG: --system-reserved-cgroup="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951791 4714 flags.go:64] FLAG: --tls-cert-file="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951800 4714 flags.go:64] FLAG: --tls-cipher-suites="[]" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951811 4714 flags.go:64] FLAG: --tls-min-version="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951819 4714 flags.go:64] FLAG: --tls-private-key-file="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951828 4714 flags.go:64] FLAG: --topology-manager-policy="none" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951837 4714 flags.go:64] FLAG: --topology-manager-policy-options="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951846 4714 flags.go:64] FLAG: --topology-manager-scope="container" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951855 4714 flags.go:64] FLAG: --v="2" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951867 4714 flags.go:64] FLAG: --version="false" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951878 4714 flags.go:64] FLAG: --vmodule="" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951889 4714 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.951898 4714 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952138 4714 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952149 4714 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952159 4714 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952168 4714 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952176 4714 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952185 4714 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952195 4714 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952205 4714 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952214 4714 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952224 4714 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952234 4714 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952243 4714 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952252 4714 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952260 4714 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952268 4714 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952277 4714 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952286 4714 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952295 4714 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952330 4714 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952341 4714 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952350 4714 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952359 4714 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952367 4714 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952374 4714 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952382 4714 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952391 4714 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952399 4714 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952407 4714 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952415 4714 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952425 4714 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952434 4714 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952444 4714 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952453 4714 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952461 4714 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952469 4714 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952478 4714 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952486 4714 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952497 4714 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952507 4714 feature_gate.go:330] unrecognized feature gate: Example Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952515 4714 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952523 4714 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952531 4714 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952539 4714 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952546 4714 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952554 4714 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952562 4714 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952569 4714 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952578 4714 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952586 4714 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952594 4714 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952602 4714 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952610 4714 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952617 4714 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952625 4714 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952633 4714 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952640 4714 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952648 4714 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952656 4714 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952664 4714 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952672 4714 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952681 4714 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952689 4714 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952697 4714 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952705 4714 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952713 4714 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952720 4714 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952728 4714 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952736 4714 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952743 4714 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952751 4714 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.952759 4714 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.955674 4714 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.968277 4714 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.968330 4714 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968490 4714 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968510 4714 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968522 4714 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968533 4714 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968544 4714 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968554 4714 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968564 4714 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968574 4714 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968585 4714 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968595 4714 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968605 4714 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968615 4714 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968626 4714 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968636 4714 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968646 4714 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968655 4714 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968666 4714 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968675 4714 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968686 4714 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968696 4714 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968706 4714 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968716 4714 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968726 4714 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968736 4714 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968747 4714 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968758 4714 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968767 4714 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968777 4714 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968786 4714 feature_gate.go:330] unrecognized feature gate: Example Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968797 4714 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968807 4714 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968817 4714 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968827 4714 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968837 4714 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968850 4714 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968860 4714 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968871 4714 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968882 4714 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968892 4714 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968902 4714 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968912 4714 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968922 4714 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968966 4714 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968977 4714 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.968987 4714 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969001 4714 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969016 4714 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969027 4714 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969039 4714 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969051 4714 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969062 4714 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969072 4714 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969086 4714 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969100 4714 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969114 4714 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969129 4714 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969140 4714 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969150 4714 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969161 4714 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969171 4714 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969185 4714 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969199 4714 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969212 4714 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969224 4714 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969236 4714 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969247 4714 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969258 4714 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969268 4714 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969279 4714 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969290 4714 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969304 4714 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.969324 4714 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969628 4714 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969649 4714 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969661 4714 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969672 4714 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969682 4714 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969692 4714 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969702 4714 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969713 4714 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969724 4714 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969734 4714 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969748 4714 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969764 4714 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969775 4714 feature_gate.go:330] unrecognized feature gate: Example Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969786 4714 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969797 4714 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969807 4714 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969817 4714 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969827 4714 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969837 4714 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969848 4714 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969857 4714 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969868 4714 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969878 4714 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969888 4714 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969899 4714 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969909 4714 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969919 4714 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969962 4714 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969974 4714 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969985 4714 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.969996 4714 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970007 4714 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970017 4714 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970027 4714 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970039 4714 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970050 4714 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970064 4714 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970078 4714 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970089 4714 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970099 4714 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970109 4714 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970119 4714 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970130 4714 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970140 4714 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970149 4714 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970160 4714 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970170 4714 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970180 4714 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970190 4714 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970199 4714 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970210 4714 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970220 4714 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970230 4714 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970240 4714 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970250 4714 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970261 4714 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970273 4714 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970283 4714 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970293 4714 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970303 4714 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970314 4714 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970325 4714 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970335 4714 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970348 4714 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970362 4714 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970373 4714 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970386 4714 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970399 4714 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970409 4714 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970419 4714 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 29 16:09:53 crc kubenswrapper[4714]: W0129 16:09:53.970431 4714 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.970447 4714 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.971570 4714 server.go:940] "Client rotation is on, will bootstrap in background" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.978720 4714 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.978915 4714 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.980970 4714 server.go:997] "Starting client certificate rotation" Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.981024 4714 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.982208 4714 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-22 23:29:39.189693331 +0000 UTC Jan 29 16:09:53 crc kubenswrapper[4714]: I0129 16:09:53.982358 4714 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.009868 4714 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 29 16:09:54 crc kubenswrapper[4714]: E0129 16:09:54.012657 4714 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.016546 4714 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.037807 4714 log.go:25] "Validated CRI v1 runtime API" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.079814 4714 log.go:25] "Validated CRI v1 image API" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.084400 4714 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.090222 4714 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-01-29-16-05-32-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.090258 4714 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.104584 4714 manager.go:217] Machine: {Timestamp:2026-01-29 16:09:54.102181766 +0000 UTC m=+0.622682896 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:1ab8f43b-7f84-4fd2-a80a-2aae14146bf4 BootID:856e4040-197b-4e74-9239-c0ebcf6976ae Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:44:c3:66 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:44:c3:66 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:5f:e7:32 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:ca:2c:4b Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:5b:93:26 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:e9:13:2b Speed:-1 Mtu:1496} {Name:eth10 MacAddress:4e:91:6c:ac:7a:42 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:d6:cd:d5:d7:8c:20 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.104786 4714 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.104877 4714 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.105150 4714 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.105287 4714 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.105312 4714 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.105479 4714 topology_manager.go:138] "Creating topology manager with none policy" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.105489 4714 container_manager_linux.go:303] "Creating device plugin manager" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.105914 4714 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.105952 4714 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.106071 4714 state_mem.go:36] "Initialized new in-memory state store" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.106146 4714 server.go:1245] "Using root directory" path="/var/lib/kubelet" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.110481 4714 kubelet.go:418] "Attempting to sync node with API server" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.110501 4714 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.110521 4714 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.110532 4714 kubelet.go:324] "Adding apiserver pod source" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.110542 4714 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.114579 4714 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.115852 4714 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.117671 4714 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 29 16:09:54 crc kubenswrapper[4714]: W0129 16:09:54.118583 4714 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Jan 29 16:09:54 crc kubenswrapper[4714]: E0129 16:09:54.118643 4714 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Jan 29 16:09:54 crc kubenswrapper[4714]: W0129 16:09:54.118718 4714 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Jan 29 16:09:54 crc kubenswrapper[4714]: E0129 16:09:54.118813 4714 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.118971 4714 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.118994 4714 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.119001 4714 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.119018 4714 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.119029 4714 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.119036 4714 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.119066 4714 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.119078 4714 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.119087 4714 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.119095 4714 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.119114 4714 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.119121 4714 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.119732 4714 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.120218 4714 server.go:1280] "Started kubelet" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.120849 4714 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.121203 4714 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 29 16:09:54 crc systemd[1]: Started Kubernetes Kubelet. Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.121880 4714 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.122531 4714 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.123880 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.123977 4714 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.124032 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 23:07:12.104438175 +0000 UTC Jan 29 16:09:54 crc kubenswrapper[4714]: E0129 16:09:54.124138 4714 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.124267 4714 volume_manager.go:287] "The desired_state_of_world populator starts" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.124287 4714 volume_manager.go:289] "Starting Kubelet Volume Manager" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.124398 4714 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 29 16:09:54 crc kubenswrapper[4714]: E0129 16:09:54.124552 4714 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="200ms" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.124997 4714 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.125015 4714 factory.go:55] Registering systemd factory Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.125024 4714 factory.go:221] Registration of the systemd container factory successfully Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.125054 4714 server.go:460] "Adding debug handlers to kubelet server" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.125700 4714 factory.go:153] Registering CRI-O factory Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.125738 4714 factory.go:221] Registration of the crio container factory successfully Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.125772 4714 factory.go:103] Registering Raw factory Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.125794 4714 manager.go:1196] Started watching for new ooms in manager Jan 29 16:09:54 crc kubenswrapper[4714]: W0129 16:09:54.126057 4714 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Jan 29 16:09:54 crc kubenswrapper[4714]: E0129 16:09:54.126179 4714 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.129087 4714 manager.go:319] Starting recovery of all containers Jan 29 16:09:54 crc kubenswrapper[4714]: E0129 16:09:54.137522 4714 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.46:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188f3f7f327a95fb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-29 16:09:54.120185339 +0000 UTC m=+0.640686459,LastTimestamp:2026-01-29 16:09:54.120185339 +0000 UTC m=+0.640686459,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.143364 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.143455 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.143488 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.143515 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.143543 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.143570 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.143595 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.143621 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.143650 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.143680 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.143708 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.143734 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.143759 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.143791 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.143817 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.143846 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.143873 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.143902 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.143928 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144085 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144108 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144130 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144151 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144173 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144193 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144213 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144238 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144293 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144372 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144399 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144426 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144453 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144483 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144509 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144535 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144566 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144594 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144624 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144650 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144677 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144702 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144727 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144753 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144780 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144807 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144874 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144906 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.144968 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145004 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145032 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145059 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145087 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145126 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145160 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145189 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145218 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145245 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145274 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145302 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145328 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145357 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145384 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145409 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145436 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145467 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145506 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145537 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145563 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145589 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145616 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145642 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145667 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145692 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145719 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145747 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145773 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145799 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145826 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145853 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145879 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145922 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.145993 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.146020 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.146045 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.146074 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.146105 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.146133 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.146158 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.147960 4714 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148275 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148294 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148309 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148323 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148336 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148353 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148369 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148384 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148413 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148428 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148448 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148461 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148476 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148490 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148501 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148517 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148541 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148558 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148575 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148590 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148603 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148617 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148630 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148647 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148663 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148677 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148692 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148706 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148717 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148734 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148748 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148764 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148778 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148791 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148805 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148822 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148837 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148850 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148864 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148878 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148889 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148952 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148969 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148980 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.148991 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149004 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149017 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149030 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149041 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149057 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149069 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149082 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149095 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149108 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149120 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149142 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149157 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149171 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149185 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149198 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149211 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149225 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149236 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149250 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149264 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149276 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149288 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149302 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149315 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149327 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149337 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149351 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149365 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149376 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149387 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149397 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149407 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149422 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149434 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149446 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149460 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149472 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149484 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149496 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149508 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149520 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149533 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149547 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149557 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149569 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149582 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149594 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149607 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149619 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149631 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149644 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149655 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149667 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149678 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149692 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149705 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149716 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149729 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149741 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149753 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149766 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149778 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149821 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149835 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149882 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149954 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149978 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.149996 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.150017 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.150034 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.150050 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.150066 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.150082 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.150097 4714 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.150111 4714 reconstruct.go:97] "Volume reconstruction finished" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.150122 4714 reconciler.go:26] "Reconciler: start to sync state" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.154332 4714 manager.go:324] Recovery completed Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.166195 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.170006 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.170050 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.170067 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.171678 4714 cpu_manager.go:225] "Starting CPU manager" policy="none" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.171701 4714 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.171723 4714 state_mem.go:36] "Initialized new in-memory state store" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.180743 4714 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.182858 4714 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.182896 4714 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.182923 4714 kubelet.go:2335] "Starting kubelet main sync loop" Jan 29 16:09:54 crc kubenswrapper[4714]: E0129 16:09:54.182986 4714 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 29 16:09:54 crc kubenswrapper[4714]: W0129 16:09:54.187156 4714 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Jan 29 16:09:54 crc kubenswrapper[4714]: E0129 16:09:54.187252 4714 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.190253 4714 policy_none.go:49] "None policy: Start" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.190959 4714 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.190984 4714 state_mem.go:35] "Initializing new in-memory state store" Jan 29 16:09:54 crc kubenswrapper[4714]: E0129 16:09:54.224477 4714 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.252396 4714 manager.go:334] "Starting Device Plugin manager" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.252669 4714 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.252698 4714 server.go:79] "Starting device plugin registration server" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.253195 4714 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.253222 4714 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.253390 4714 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.253528 4714 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.253557 4714 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 29 16:09:54 crc kubenswrapper[4714]: E0129 16:09:54.262526 4714 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.283886 4714 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.284012 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.285301 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.285339 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.285349 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.285481 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.285675 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.285717 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.286243 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.286266 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.286274 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.286652 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.286693 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.286706 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.286850 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.286991 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.287030 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.287618 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.287665 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.287673 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.287738 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.287755 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.287849 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.287755 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.287885 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.287905 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.288424 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.288449 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.288457 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.288541 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.288616 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.288660 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.289144 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.289167 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.289175 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.289287 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.289307 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.289731 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.289752 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.289760 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.289796 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.289812 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.289822 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.289910 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.289926 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.289961 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:54 crc kubenswrapper[4714]: E0129 16:09:54.324988 4714 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="400ms" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.353520 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.353581 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.353630 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.353672 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.353731 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.353763 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.353780 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.353804 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.353922 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.353971 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.354009 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.354033 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.354056 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.354092 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.354118 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.354172 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.354436 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.354462 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.354469 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.354490 4714 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 16:09:54 crc kubenswrapper[4714]: E0129 16:09:54.354968 4714 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.46:6443: connect: connection refused" node="crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.454790 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.454872 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.454908 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.454987 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455006 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455059 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455087 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455092 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455085 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455019 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455175 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455204 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455232 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455260 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455290 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455333 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455378 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455397 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455424 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455378 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455446 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455461 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455475 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455477 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455475 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455484 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455420 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455623 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455534 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.455587 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.555872 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.557320 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.557368 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.557386 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.557417 4714 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 16:09:54 crc kubenswrapper[4714]: E0129 16:09:54.558012 4714 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.46:6443: connect: connection refused" node="crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.623774 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.647152 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.654347 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: W0129 16:09:54.663095 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-593d0f6b670c6966a3bbaadaefb40a1ad42f1fa8edf9796bf0b8d0301dc388ac WatchSource:0}: Error finding container 593d0f6b670c6966a3bbaadaefb40a1ad42f1fa8edf9796bf0b8d0301dc388ac: Status 404 returned error can't find the container with id 593d0f6b670c6966a3bbaadaefb40a1ad42f1fa8edf9796bf0b8d0301dc388ac Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.672979 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: W0129 16:09:54.674060 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-999f1dcb3adbb51cdabe6027fe9de8017a8a05762d46b433c8d9e7661a1bbc43 WatchSource:0}: Error finding container 999f1dcb3adbb51cdabe6027fe9de8017a8a05762d46b433c8d9e7661a1bbc43: Status 404 returned error can't find the container with id 999f1dcb3adbb51cdabe6027fe9de8017a8a05762d46b433c8d9e7661a1bbc43 Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.679188 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 16:09:54 crc kubenswrapper[4714]: W0129 16:09:54.692003 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-68be8e1eb6567776239bcd622bc35f2239a081987843d4b6a00afa70ddd507c6 WatchSource:0}: Error finding container 68be8e1eb6567776239bcd622bc35f2239a081987843d4b6a00afa70ddd507c6: Status 404 returned error can't find the container with id 68be8e1eb6567776239bcd622bc35f2239a081987843d4b6a00afa70ddd507c6 Jan 29 16:09:54 crc kubenswrapper[4714]: W0129 16:09:54.701719 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-239eb8666b3fe82cedcc35d13f0a5d821d908c81ea42456e4d6f8a03ec8fd490 WatchSource:0}: Error finding container 239eb8666b3fe82cedcc35d13f0a5d821d908c81ea42456e4d6f8a03ec8fd490: Status 404 returned error can't find the container with id 239eb8666b3fe82cedcc35d13f0a5d821d908c81ea42456e4d6f8a03ec8fd490 Jan 29 16:09:54 crc kubenswrapper[4714]: E0129 16:09:54.725880 4714 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="800ms" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.959084 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.960291 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.960348 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.960360 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:54 crc kubenswrapper[4714]: I0129 16:09:54.960394 4714 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 16:09:54 crc kubenswrapper[4714]: E0129 16:09:54.960837 4714 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.46:6443: connect: connection refused" node="crc" Jan 29 16:09:54 crc kubenswrapper[4714]: W0129 16:09:54.964590 4714 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Jan 29 16:09:54 crc kubenswrapper[4714]: E0129 16:09:54.964666 4714 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Jan 29 16:09:55 crc kubenswrapper[4714]: W0129 16:09:55.047554 4714 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Jan 29 16:09:55 crc kubenswrapper[4714]: E0129 16:09:55.047640 4714 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Jan 29 16:09:55 crc kubenswrapper[4714]: I0129 16:09:55.121625 4714 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Jan 29 16:09:55 crc kubenswrapper[4714]: I0129 16:09:55.124723 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 14:54:20.541537821 +0000 UTC Jan 29 16:09:55 crc kubenswrapper[4714]: I0129 16:09:55.186335 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"239eb8666b3fe82cedcc35d13f0a5d821d908c81ea42456e4d6f8a03ec8fd490"} Jan 29 16:09:55 crc kubenswrapper[4714]: I0129 16:09:55.186971 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"68be8e1eb6567776239bcd622bc35f2239a081987843d4b6a00afa70ddd507c6"} Jan 29 16:09:55 crc kubenswrapper[4714]: I0129 16:09:55.187718 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a1d20cbc534d7ed08405bfd39639adc94428bd1e01a3e749031288f444f59d07"} Jan 29 16:09:55 crc kubenswrapper[4714]: I0129 16:09:55.188375 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"999f1dcb3adbb51cdabe6027fe9de8017a8a05762d46b433c8d9e7661a1bbc43"} Jan 29 16:09:55 crc kubenswrapper[4714]: I0129 16:09:55.189148 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"593d0f6b670c6966a3bbaadaefb40a1ad42f1fa8edf9796bf0b8d0301dc388ac"} Jan 29 16:09:55 crc kubenswrapper[4714]: W0129 16:09:55.348117 4714 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Jan 29 16:09:55 crc kubenswrapper[4714]: E0129 16:09:55.348211 4714 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Jan 29 16:09:55 crc kubenswrapper[4714]: W0129 16:09:55.397269 4714 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Jan 29 16:09:55 crc kubenswrapper[4714]: E0129 16:09:55.397349 4714 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Jan 29 16:09:55 crc kubenswrapper[4714]: E0129 16:09:55.526713 4714 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="1.6s" Jan 29 16:09:55 crc kubenswrapper[4714]: I0129 16:09:55.761099 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:55 crc kubenswrapper[4714]: I0129 16:09:55.762140 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:55 crc kubenswrapper[4714]: I0129 16:09:55.762184 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:55 crc kubenswrapper[4714]: I0129 16:09:55.762193 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:55 crc kubenswrapper[4714]: I0129 16:09:55.762224 4714 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 16:09:55 crc kubenswrapper[4714]: E0129 16:09:55.762674 4714 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.46:6443: connect: connection refused" node="crc" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.075805 4714 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 29 16:09:56 crc kubenswrapper[4714]: E0129 16:09:56.076860 4714 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.122601 4714 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.124999 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 05:29:21.169865698 +0000 UTC Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.193523 4714 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c" exitCode=0 Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.193577 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c"} Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.193758 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.194946 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.194981 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.194993 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.196425 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.196955 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995"} Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.196985 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30"} Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.196993 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.196998 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2"} Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.197008 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835"} Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.197309 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.197337 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.197347 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.197996 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.198076 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.198097 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.199956 4714 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7" exitCode=0 Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.200006 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7"} Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.200149 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.201295 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.201328 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.201340 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.201367 4714 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="86530d623401c3cf5ebe44fcc1c2110a5f4dde059a5d27f74e118d5ca1df40dc" exitCode=0 Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.201427 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"86530d623401c3cf5ebe44fcc1c2110a5f4dde059a5d27f74e118d5ca1df40dc"} Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.201523 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.202410 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.202451 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.202466 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.203582 4714 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec" exitCode=0 Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.203627 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec"} Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.203659 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.204335 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.204359 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:56 crc kubenswrapper[4714]: I0129 16:09:56.204367 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:56 crc kubenswrapper[4714]: W0129 16:09:56.587847 4714 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Jan 29 16:09:56 crc kubenswrapper[4714]: E0129 16:09:56.587974 4714 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Jan 29 16:09:57 crc kubenswrapper[4714]: W0129 16:09:57.104463 4714 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Jan 29 16:09:57 crc kubenswrapper[4714]: E0129 16:09:57.104546 4714 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.122088 4714 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.125402 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 18:08:51.746540599 +0000 UTC Jan 29 16:09:57 crc kubenswrapper[4714]: E0129 16:09:57.128156 4714 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="3.2s" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.207047 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6565dc4c35c618a239bdc2be6dd45a9057e83573c92c3fc816350eb014c822c9"} Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.207085 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.207091 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6068173079b1d5cf7de92fe34bd0a4701b11a4dba7ae384f2d0e33f185656107"} Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.207192 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1f0bdc60ea5a5e188b2ed8894bb387084567b294c3a356fed01493d4bd6a7caf"} Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.208138 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.208212 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.208265 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.211538 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2"} Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.211629 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa"} Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.211732 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6"} Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.211791 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f"} Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.213087 4714 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c" exitCode=0 Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.213204 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c"} Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.213366 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.214023 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.214102 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.214165 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.215966 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.216371 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.216469 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"89cb2f8d441042c4c95e3cc056f991565c18bd93dcb0d61f3735e2451ff439a4"} Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.217136 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.217244 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.217283 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.217376 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.217404 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.217327 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.363250 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.364361 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.364389 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.364398 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:57 crc kubenswrapper[4714]: I0129 16:09:57.364419 4714 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 16:09:57 crc kubenswrapper[4714]: E0129 16:09:57.364811 4714 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.46:6443: connect: connection refused" node="crc" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.125632 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 01:08:43.355338261 +0000 UTC Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.222052 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8"} Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.222567 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.223974 4714 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334" exitCode=0 Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.224107 4714 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.224162 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.224372 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.224537 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.224701 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.224948 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.224992 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.224857 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334"} Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.225911 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.226046 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.226057 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.226068 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.226078 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.226075 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.226177 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.226173 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.226192 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.512668 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.512852 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.514484 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.514523 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.514535 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:58 crc kubenswrapper[4714]: I0129 16:09:58.521418 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:09:59 crc kubenswrapper[4714]: I0129 16:09:59.126742 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 17:16:10.736502391 +0000 UTC Jan 29 16:09:59 crc kubenswrapper[4714]: I0129 16:09:59.233620 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:59 crc kubenswrapper[4714]: I0129 16:09:59.233595 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346"} Jan 29 16:09:59 crc kubenswrapper[4714]: I0129 16:09:59.233799 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:09:59 crc kubenswrapper[4714]: I0129 16:09:59.233804 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9"} Jan 29 16:09:59 crc kubenswrapper[4714]: I0129 16:09:59.233948 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:09:59 crc kubenswrapper[4714]: I0129 16:09:59.233974 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02"} Jan 29 16:09:59 crc kubenswrapper[4714]: I0129 16:09:59.233992 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109"} Jan 29 16:09:59 crc kubenswrapper[4714]: I0129 16:09:59.235178 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:59 crc kubenswrapper[4714]: I0129 16:09:59.235209 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:09:59 crc kubenswrapper[4714]: I0129 16:09:59.235248 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:59 crc kubenswrapper[4714]: I0129 16:09:59.235266 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:09:59 crc kubenswrapper[4714]: I0129 16:09:59.235224 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:09:59 crc kubenswrapper[4714]: I0129 16:09:59.235331 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.127255 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 23:25:20.672932191 +0000 UTC Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.177509 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.180683 4714 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.240173 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4"} Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.240246 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.240278 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.240248 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.241337 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.241359 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.241369 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.241501 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.241536 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.241550 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.241705 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.241811 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.241904 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.565608 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.567075 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.567167 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.567231 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:00 crc kubenswrapper[4714]: I0129 16:10:00.567310 4714 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.127538 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 13:30:41.544701161 +0000 UTC Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.243335 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.244335 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.244465 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.244580 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.393670 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.394337 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.396279 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.396333 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.396355 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.450512 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.450734 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.452402 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.452443 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.452456 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.886056 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.886343 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.887726 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.887767 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.887776 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:01 crc kubenswrapper[4714]: I0129 16:10:01.995987 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:10:02 crc kubenswrapper[4714]: I0129 16:10:02.128199 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 18:14:30.978735463 +0000 UTC Jan 29 16:10:02 crc kubenswrapper[4714]: I0129 16:10:02.246409 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:10:02 crc kubenswrapper[4714]: I0129 16:10:02.248307 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:02 crc kubenswrapper[4714]: I0129 16:10:02.248358 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:02 crc kubenswrapper[4714]: I0129 16:10:02.248374 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:02 crc kubenswrapper[4714]: I0129 16:10:02.373874 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Jan 29 16:10:02 crc kubenswrapper[4714]: I0129 16:10:02.374138 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:10:02 crc kubenswrapper[4714]: I0129 16:10:02.375841 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:02 crc kubenswrapper[4714]: I0129 16:10:02.375913 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:02 crc kubenswrapper[4714]: I0129 16:10:02.375973 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:03 crc kubenswrapper[4714]: I0129 16:10:03.128524 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 16:58:38.912384011 +0000 UTC Jan 29 16:10:04 crc kubenswrapper[4714]: I0129 16:10:04.129264 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 21:25:48.803572412 +0000 UTC Jan 29 16:10:04 crc kubenswrapper[4714]: E0129 16:10:04.262659 4714 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 29 16:10:04 crc kubenswrapper[4714]: I0129 16:10:04.735399 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:10:04 crc kubenswrapper[4714]: I0129 16:10:04.735558 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:10:04 crc kubenswrapper[4714]: I0129 16:10:04.736701 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:04 crc kubenswrapper[4714]: I0129 16:10:04.736741 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:04 crc kubenswrapper[4714]: I0129 16:10:04.736759 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:05 crc kubenswrapper[4714]: I0129 16:10:05.129629 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 16:23:47.839923433 +0000 UTC Jan 29 16:10:06 crc kubenswrapper[4714]: I0129 16:10:06.129998 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 00:25:23.73045319 +0000 UTC Jan 29 16:10:06 crc kubenswrapper[4714]: I0129 16:10:06.441560 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Jan 29 16:10:06 crc kubenswrapper[4714]: I0129 16:10:06.441714 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:10:06 crc kubenswrapper[4714]: I0129 16:10:06.442975 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:06 crc kubenswrapper[4714]: I0129 16:10:06.443000 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:06 crc kubenswrapper[4714]: I0129 16:10:06.443008 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:07 crc kubenswrapper[4714]: I0129 16:10:07.130440 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 04:39:08.67083275 +0000 UTC Jan 29 16:10:07 crc kubenswrapper[4714]: I0129 16:10:07.736235 4714 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 29 16:10:07 crc kubenswrapper[4714]: I0129 16:10:07.736339 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 29 16:10:08 crc kubenswrapper[4714]: I0129 16:10:08.122417 4714 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Jan 29 16:10:08 crc kubenswrapper[4714]: I0129 16:10:08.130902 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 19:57:14.757433164 +0000 UTC Jan 29 16:10:08 crc kubenswrapper[4714]: W0129 16:10:08.174275 4714 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Jan 29 16:10:08 crc kubenswrapper[4714]: I0129 16:10:08.174384 4714 trace.go:236] Trace[1706721841]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Jan-2026 16:09:58.172) (total time: 10001ms): Jan 29 16:10:08 crc kubenswrapper[4714]: Trace[1706721841]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (16:10:08.174) Jan 29 16:10:08 crc kubenswrapper[4714]: Trace[1706721841]: [10.001636696s] [10.001636696s] END Jan 29 16:10:08 crc kubenswrapper[4714]: E0129 16:10:08.174431 4714 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Jan 29 16:10:08 crc kubenswrapper[4714]: I0129 16:10:08.240368 4714 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46538->192.168.126.11:17697: read: connection reset by peer" start-of-body= Jan 29 16:10:08 crc kubenswrapper[4714]: I0129 16:10:08.240464 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46538->192.168.126.11:17697: read: connection reset by peer" Jan 29 16:10:08 crc kubenswrapper[4714]: W0129 16:10:08.508717 4714 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Jan 29 16:10:08 crc kubenswrapper[4714]: I0129 16:10:08.508851 4714 trace.go:236] Trace[1900129430]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Jan-2026 16:09:58.507) (total time: 10001ms): Jan 29 16:10:08 crc kubenswrapper[4714]: Trace[1900129430]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (16:10:08.508) Jan 29 16:10:08 crc kubenswrapper[4714]: Trace[1900129430]: [10.001248415s] [10.001248415s] END Jan 29 16:10:08 crc kubenswrapper[4714]: E0129 16:10:08.508903 4714 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Jan 29 16:10:09 crc kubenswrapper[4714]: I0129 16:10:09.131391 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 02:11:31.172349292 +0000 UTC Jan 29 16:10:09 crc kubenswrapper[4714]: I0129 16:10:09.267394 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 29 16:10:09 crc kubenswrapper[4714]: I0129 16:10:09.270604 4714 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8" exitCode=255 Jan 29 16:10:09 crc kubenswrapper[4714]: I0129 16:10:09.270667 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8"} Jan 29 16:10:09 crc kubenswrapper[4714]: I0129 16:10:09.270915 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:10:09 crc kubenswrapper[4714]: I0129 16:10:09.272259 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:09 crc kubenswrapper[4714]: I0129 16:10:09.272336 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:09 crc kubenswrapper[4714]: I0129 16:10:09.272356 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:09 crc kubenswrapper[4714]: I0129 16:10:09.273402 4714 scope.go:117] "RemoveContainer" containerID="6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8" Jan 29 16:10:09 crc kubenswrapper[4714]: I0129 16:10:09.373467 4714 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 29 16:10:09 crc kubenswrapper[4714]: I0129 16:10:09.373540 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 29 16:10:09 crc kubenswrapper[4714]: I0129 16:10:09.379763 4714 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 29 16:10:09 crc kubenswrapper[4714]: I0129 16:10:09.379828 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 29 16:10:10 crc kubenswrapper[4714]: I0129 16:10:10.131858 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 02:09:58.935019265 +0000 UTC Jan 29 16:10:10 crc kubenswrapper[4714]: I0129 16:10:10.184500 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:10:10 crc kubenswrapper[4714]: I0129 16:10:10.184706 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:10:10 crc kubenswrapper[4714]: I0129 16:10:10.186214 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:10 crc kubenswrapper[4714]: I0129 16:10:10.186276 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:10 crc kubenswrapper[4714]: I0129 16:10:10.186301 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:10 crc kubenswrapper[4714]: I0129 16:10:10.279334 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 29 16:10:10 crc kubenswrapper[4714]: I0129 16:10:10.282468 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79"} Jan 29 16:10:10 crc kubenswrapper[4714]: I0129 16:10:10.282708 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:10:10 crc kubenswrapper[4714]: I0129 16:10:10.283873 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:10 crc kubenswrapper[4714]: I0129 16:10:10.283949 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:10 crc kubenswrapper[4714]: I0129 16:10:10.283975 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:11 crc kubenswrapper[4714]: I0129 16:10:11.132694 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 11:45:14.889820284 +0000 UTC Jan 29 16:10:12 crc kubenswrapper[4714]: I0129 16:10:12.003538 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:10:12 crc kubenswrapper[4714]: I0129 16:10:12.003817 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:10:12 crc kubenswrapper[4714]: I0129 16:10:12.003988 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:10:12 crc kubenswrapper[4714]: I0129 16:10:12.005519 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:12 crc kubenswrapper[4714]: I0129 16:10:12.005578 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:12 crc kubenswrapper[4714]: I0129 16:10:12.005598 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:12 crc kubenswrapper[4714]: I0129 16:10:12.013150 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:10:12 crc kubenswrapper[4714]: I0129 16:10:12.133240 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 06:55:36.331551379 +0000 UTC Jan 29 16:10:12 crc kubenswrapper[4714]: I0129 16:10:12.288204 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:10:12 crc kubenswrapper[4714]: I0129 16:10:12.289084 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:12 crc kubenswrapper[4714]: I0129 16:10:12.289127 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:12 crc kubenswrapper[4714]: I0129 16:10:12.289145 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:12 crc kubenswrapper[4714]: I0129 16:10:12.581157 4714 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.121895 4714 apiserver.go:52] "Watching apiserver" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.133738 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 22:55:27.025353911 +0000 UTC Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.139409 4714 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.139901 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.140505 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.140690 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:13 crc kubenswrapper[4714]: E0129 16:10:13.140870 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.141037 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.141048 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.141244 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:13 crc kubenswrapper[4714]: E0129 16:10:13.141289 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.141681 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:13 crc kubenswrapper[4714]: E0129 16:10:13.141770 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.142978 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.144208 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.144422 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.144457 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.145049 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.145594 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.145838 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.145877 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.146308 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.173663 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.194594 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.207695 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.221275 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.225827 4714 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.232647 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.244586 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.261950 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:13 crc kubenswrapper[4714]: I0129 16:10:13.308185 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.003334 4714 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.134974 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 03:57:01.551416911 +0000 UTC Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.204258 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.216913 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.227996 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.239116 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.249755 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.266066 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.277096 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: E0129 16:10:14.373810 4714 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Jan 29 16:10:14 crc kubenswrapper[4714]: E0129 16:10:14.376463 4714 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.376733 4714 trace.go:236] Trace[2076281891]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Jan-2026 16:10:01.075) (total time: 13301ms): Jan 29 16:10:14 crc kubenswrapper[4714]: Trace[2076281891]: ---"Objects listed" error: 13301ms (16:10:14.376) Jan 29 16:10:14 crc kubenswrapper[4714]: Trace[2076281891]: [13.301400826s] [13.301400826s] END Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.377171 4714 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.378599 4714 trace.go:236] Trace[298477415]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Jan-2026 16:10:01.047) (total time: 13331ms): Jan 29 16:10:14 crc kubenswrapper[4714]: Trace[298477415]: ---"Objects listed" error: 13331ms (16:10:14.378) Jan 29 16:10:14 crc kubenswrapper[4714]: Trace[298477415]: [13.3310769s] [13.3310769s] END Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.378900 4714 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.378658 4714 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.387959 4714 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.410702 4714 csr.go:261] certificate signing request csr-2pk7r is approved, waiting to be issued Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.420149 4714 csr.go:257] certificate signing request csr-2pk7r is issued Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.479734 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.480268 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.480470 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.481263 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.482794 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.483445 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.484398 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.485025 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.485485 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.485590 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.486098 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.486209 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.486311 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.486401 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.480738 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.486494 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.486486 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.480746 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.481177 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.486601 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.482707 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.483355 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.486638 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.486670 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.486695 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.486721 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.486745 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.484320 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.484971 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.485418 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.486155 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.486876 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.486884 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.486907 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.487401 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.487499 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.487542 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.487614 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.487110 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.487129 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.487146 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.487173 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.487295 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.487438 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.487439 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.487466 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.487744 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.487796 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.487813 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.487839 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.487824 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.487863 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.487952 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.487979 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488001 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488035 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488066 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488087 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488097 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488131 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488160 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488165 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488191 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488226 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488258 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488288 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488306 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488319 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488328 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488351 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488410 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488471 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488508 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488542 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488566 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488576 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488614 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488650 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488683 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488716 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488752 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488782 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488814 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488861 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488618 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488896 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488954 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488994 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489030 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489066 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489096 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488889 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488660 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488652 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489156 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489167 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488794 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489189 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489205 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488832 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488877 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488962 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.488990 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489096 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489355 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489407 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489387 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489423 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489432 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489598 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489618 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489714 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489748 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489134 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489774 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489808 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489839 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489863 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489886 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.489909 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.490093 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.490701 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.490800 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.490843 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.490882 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.490928 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.490998 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491039 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491080 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491117 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491169 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491205 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491242 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491248 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491281 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491319 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491361 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491400 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491439 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491476 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491514 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491548 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491586 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491749 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491787 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491826 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491860 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491870 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491910 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.491976 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492025 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492061 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492096 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492132 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492170 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492206 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492246 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492288 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492324 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492358 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492425 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492461 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492495 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492529 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492622 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492657 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492690 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492723 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492787 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492827 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492861 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492901 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492960 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493005 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493042 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493079 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493115 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493153 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493188 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493269 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493307 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493344 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493391 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493447 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493487 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493526 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493562 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493600 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493643 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493695 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493739 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493784 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493824 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493863 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493900 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493965 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494003 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494040 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494077 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494112 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494147 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494183 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494220 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494276 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494311 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494347 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494385 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494421 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494465 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494502 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494539 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494577 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494617 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494656 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494694 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494732 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494773 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494814 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494853 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494889 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494927 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494986 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495031 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495067 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495102 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495139 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495172 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495208 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495243 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495278 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495313 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495349 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495384 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495422 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495458 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495493 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495535 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495576 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495613 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495651 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495687 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495727 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495764 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495800 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495836 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495878 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495918 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495987 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.496032 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.496069 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.496103 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.496204 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.496273 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.496367 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.496423 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.496473 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.496530 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.496570 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.496615 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.496659 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.496697 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.496749 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.496787 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.496833 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.496877 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501105 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501168 4714 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501206 4714 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501237 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501271 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501301 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501323 4714 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501348 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501377 4714 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501407 4714 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501447 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501472 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501493 4714 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501513 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501534 4714 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501561 4714 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501590 4714 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501617 4714 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501645 4714 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501673 4714 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501693 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501713 4714 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501736 4714 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501755 4714 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501794 4714 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501819 4714 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501838 4714 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501859 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501881 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501903 4714 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501924 4714 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501974 4714 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501996 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502017 4714 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502036 4714 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502061 4714 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502091 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502122 4714 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502150 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502174 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502197 4714 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502227 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502256 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502281 4714 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502304 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502328 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502359 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502384 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502404 4714 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502423 4714 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502444 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502464 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502483 4714 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502503 4714 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502522 4714 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.502542 4714 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.492529 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.493633 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494129 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494408 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.494862 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.495216 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501107 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.501427 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.503922 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.505341 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.506349 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.507004 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.507568 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.507754 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.507828 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.508056 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.508246 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.508320 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.508434 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.508606 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.508876 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: E0129 16:10:14.508908 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:10:15.008886106 +0000 UTC m=+21.529387226 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.510809 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.510921 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.514377 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.517337 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.517353 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.517620 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.517622 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.517973 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.518201 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.518192 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.518289 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.518477 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.518631 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.518866 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.519229 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.519325 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.519553 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.519699 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.520285 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.520612 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.520917 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.520924 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.520961 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.511175 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.526372 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.526555 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.526726 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.528331 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.530189 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.530563 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.531590 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.531907 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.532342 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.532920 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.533125 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.533560 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.533666 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.533846 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.533871 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.533903 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.533960 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.534197 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.534285 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.534401 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.534428 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.534780 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.535276 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.535413 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.535782 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.538606 4714 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.542034 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.544309 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.544570 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.544799 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.544820 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.534518 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.547865 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.548056 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.548625 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.553560 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.554505 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.555052 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.555537 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.555531 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.555773 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.555787 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.556152 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.556217 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.556282 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.556341 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.556792 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.557872 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: E0129 16:10:14.559004 4714 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 16:10:14 crc kubenswrapper[4714]: E0129 16:10:14.559027 4714 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 16:10:14 crc kubenswrapper[4714]: E0129 16:10:14.559078 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:15.059058743 +0000 UTC m=+21.579559863 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 16:10:14 crc kubenswrapper[4714]: E0129 16:10:14.559162 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:15.059129505 +0000 UTC m=+21.579630635 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.560602 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 16:10:14 crc kubenswrapper[4714]: E0129 16:10:14.563109 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 16:10:14 crc kubenswrapper[4714]: E0129 16:10:14.563139 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 16:10:14 crc kubenswrapper[4714]: E0129 16:10:14.563160 4714 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:14 crc kubenswrapper[4714]: E0129 16:10:14.563225 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:15.063204084 +0000 UTC m=+21.583705204 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.564840 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.565557 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.565851 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.568906 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.569113 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.569250 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.569503 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.569520 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.570387 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.570437 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.570606 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.570606 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.570683 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.570451 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.570792 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.570800 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.571071 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.570957 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.574345 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.580225 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 16:10:14 crc kubenswrapper[4714]: E0129 16:10:14.580588 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 16:10:14 crc kubenswrapper[4714]: E0129 16:10:14.580608 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 16:10:14 crc kubenswrapper[4714]: E0129 16:10:14.580623 4714 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:14 crc kubenswrapper[4714]: E0129 16:10:14.580681 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:15.080658904 +0000 UTC m=+21.601160024 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.590160 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.596290 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.596745 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.601214 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.601628 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.601744 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.603952 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.603982 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604051 4714 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604068 4714 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604077 4714 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604086 4714 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604096 4714 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604104 4714 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604112 4714 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604122 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604132 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604140 4714 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604149 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604158 4714 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604166 4714 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604174 4714 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604182 4714 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604191 4714 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604200 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604208 4714 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604216 4714 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604224 4714 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604233 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604241 4714 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604219 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604249 4714 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604287 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604319 4714 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604338 4714 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604358 4714 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604371 4714 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604383 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604376 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604396 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604458 4714 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604475 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.604280 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.605762 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.605794 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.605807 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.605818 4714 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.605828 4714 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.605839 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.605850 4714 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.605862 4714 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.605873 4714 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.605882 4714 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.605893 4714 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.605903 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.605914 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.605923 4714 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.605949 4714 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.605959 4714 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.605968 4714 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.605977 4714 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.605987 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.605997 4714 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606005 4714 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606015 4714 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606031 4714 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606041 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606049 4714 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606058 4714 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606069 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606080 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606090 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606099 4714 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606109 4714 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606119 4714 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606131 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606141 4714 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606151 4714 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606161 4714 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606171 4714 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606182 4714 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606191 4714 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606199 4714 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606208 4714 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606218 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606226 4714 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606235 4714 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606245 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606253 4714 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606262 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606279 4714 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606287 4714 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606296 4714 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606304 4714 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606313 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606323 4714 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606333 4714 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606341 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606351 4714 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606361 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606370 4714 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606378 4714 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606387 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606397 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606405 4714 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606414 4714 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606424 4714 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606443 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606453 4714 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606462 4714 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606470 4714 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606479 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606488 4714 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606497 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606507 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606516 4714 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606525 4714 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606533 4714 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606542 4714 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606551 4714 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606559 4714 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606568 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606577 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606587 4714 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606597 4714 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606606 4714 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.606615 4714 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.611664 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.611978 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.612174 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.612226 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.612266 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.612591 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.612633 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.612840 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.613231 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.614396 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.614433 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.614944 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.615227 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.615736 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.615762 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.619390 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.621181 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.630449 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.633058 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.633061 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.633353 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.633426 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.647729 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.656074 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.656289 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.667149 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.677479 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.677984 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709023 4714 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709139 4714 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709472 4714 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709506 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709518 4714 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709530 4714 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709540 4714 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709549 4714 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709560 4714 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709570 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709581 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709590 4714 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709598 4714 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709608 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709626 4714 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709635 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709644 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709652 4714 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709660 4714 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709669 4714 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709677 4714 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709686 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.709696 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.744789 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.751097 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.764534 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.781183 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.798390 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.808953 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.811865 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.821188 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.832805 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.840827 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.850537 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.861865 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.876393 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.886631 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.908231 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.918868 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.928917 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:14 crc kubenswrapper[4714]: I0129 16:10:14.939775 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.012200 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:10:15 crc kubenswrapper[4714]: E0129 16:10:15.012418 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:10:16.012395443 +0000 UTC m=+22.532896553 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.113023 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.113083 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.113120 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.113149 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:15 crc kubenswrapper[4714]: E0129 16:10:15.113281 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 16:10:15 crc kubenswrapper[4714]: E0129 16:10:15.113282 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 16:10:15 crc kubenswrapper[4714]: E0129 16:10:15.113287 4714 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 16:10:15 crc kubenswrapper[4714]: E0129 16:10:15.113346 4714 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 16:10:15 crc kubenswrapper[4714]: E0129 16:10:15.113299 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 16:10:15 crc kubenswrapper[4714]: E0129 16:10:15.113447 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:16.113419686 +0000 UTC m=+22.633920806 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 16:10:15 crc kubenswrapper[4714]: E0129 16:10:15.113459 4714 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:15 crc kubenswrapper[4714]: E0129 16:10:15.113472 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:16.113463027 +0000 UTC m=+22.633964147 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 16:10:15 crc kubenswrapper[4714]: E0129 16:10:15.113524 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:16.113484628 +0000 UTC m=+22.633985768 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:15 crc kubenswrapper[4714]: E0129 16:10:15.113323 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 16:10:15 crc kubenswrapper[4714]: E0129 16:10:15.113547 4714 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:15 crc kubenswrapper[4714]: E0129 16:10:15.113579 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:16.11357028 +0000 UTC m=+22.634071420 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.135477 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 13:46:38.508414533 +0000 UTC Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.184045 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.184107 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:15 crc kubenswrapper[4714]: E0129 16:10:15.184267 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.184406 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:15 crc kubenswrapper[4714]: E0129 16:10:15.184412 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:15 crc kubenswrapper[4714]: E0129 16:10:15.184740 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.296801 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d"} Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.297251 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315"} Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.297323 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a3e46ca0032e7c46b49567152df4cab88b64467e2571d89c9955362b3c5a5f43"} Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.298031 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b"} Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.298097 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"86da9d03ae94b77bfb3371663564750484db9e0378cce10cfc3e10ab16a69bc0"} Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.298837 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a974d693556b28ba1f5a38ab8b3e14c0e5472255f8c6b3e4d29612bdb5151887"} Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.325029 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.343628 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.384453 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.413274 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.421965 4714 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-01-29 16:05:14 +0000 UTC, rotation deadline is 2026-11-17 06:34:21.676761944 +0000 UTC Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.422030 4714 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6998h24m6.254735008s for next certificate rotation Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.447589 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.467501 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.479858 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.490612 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.505108 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.519355 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.534834 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.549859 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.564393 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.581958 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.598878 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.612150 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.774341 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-46dqc"] Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.774749 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-46dqc" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.778361 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-ppngk"] Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.778655 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.779049 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.779128 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.779331 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.782407 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.782596 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.783981 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.784652 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.786468 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.799120 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.814341 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.830973 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.846503 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.864842 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.877546 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.895878 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.917567 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.920017 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f846b283-5468-4014-ba05-da5bfffa2ebc-hosts-file\") pod \"node-resolver-46dqc\" (UID: \"f846b283-5468-4014-ba05-da5bfffa2ebc\") " pod="openshift-dns/node-resolver-46dqc" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.920151 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c8c765f3-89eb-4077-8829-03e86eb0c90c-mcd-auth-proxy-config\") pod \"machine-config-daemon-ppngk\" (UID: \"c8c765f3-89eb-4077-8829-03e86eb0c90c\") " pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.920203 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c8c765f3-89eb-4077-8829-03e86eb0c90c-proxy-tls\") pod \"machine-config-daemon-ppngk\" (UID: \"c8c765f3-89eb-4077-8829-03e86eb0c90c\") " pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.920238 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsqf5\" (UniqueName: \"kubernetes.io/projected/c8c765f3-89eb-4077-8829-03e86eb0c90c-kube-api-access-bsqf5\") pod \"machine-config-daemon-ppngk\" (UID: \"c8c765f3-89eb-4077-8829-03e86eb0c90c\") " pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.920270 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbd9b\" (UniqueName: \"kubernetes.io/projected/f846b283-5468-4014-ba05-da5bfffa2ebc-kube-api-access-gbd9b\") pod \"node-resolver-46dqc\" (UID: \"f846b283-5468-4014-ba05-da5bfffa2ebc\") " pod="openshift-dns/node-resolver-46dqc" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.920300 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c8c765f3-89eb-4077-8829-03e86eb0c90c-rootfs\") pod \"machine-config-daemon-ppngk\" (UID: \"c8c765f3-89eb-4077-8829-03e86eb0c90c\") " pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.932100 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.949490 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.965118 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.979058 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:15 crc kubenswrapper[4714]: I0129 16:10:15.994781 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.012038 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.021267 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:10:16 crc kubenswrapper[4714]: E0129 16:10:16.021534 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:10:18.021494368 +0000 UTC m=+24.541995528 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.021814 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c8c765f3-89eb-4077-8829-03e86eb0c90c-proxy-tls\") pod \"machine-config-daemon-ppngk\" (UID: \"c8c765f3-89eb-4077-8829-03e86eb0c90c\") " pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.022071 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsqf5\" (UniqueName: \"kubernetes.io/projected/c8c765f3-89eb-4077-8829-03e86eb0c90c-kube-api-access-bsqf5\") pod \"machine-config-daemon-ppngk\" (UID: \"c8c765f3-89eb-4077-8829-03e86eb0c90c\") " pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.022302 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbd9b\" (UniqueName: \"kubernetes.io/projected/f846b283-5468-4014-ba05-da5bfffa2ebc-kube-api-access-gbd9b\") pod \"node-resolver-46dqc\" (UID: \"f846b283-5468-4014-ba05-da5bfffa2ebc\") " pod="openshift-dns/node-resolver-46dqc" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.022475 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c8c765f3-89eb-4077-8829-03e86eb0c90c-rootfs\") pod \"machine-config-daemon-ppngk\" (UID: \"c8c765f3-89eb-4077-8829-03e86eb0c90c\") " pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.022630 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f846b283-5468-4014-ba05-da5bfffa2ebc-hosts-file\") pod \"node-resolver-46dqc\" (UID: \"f846b283-5468-4014-ba05-da5bfffa2ebc\") " pod="openshift-dns/node-resolver-46dqc" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.022735 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f846b283-5468-4014-ba05-da5bfffa2ebc-hosts-file\") pod \"node-resolver-46dqc\" (UID: \"f846b283-5468-4014-ba05-da5bfffa2ebc\") " pod="openshift-dns/node-resolver-46dqc" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.022544 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c8c765f3-89eb-4077-8829-03e86eb0c90c-rootfs\") pod \"machine-config-daemon-ppngk\" (UID: \"c8c765f3-89eb-4077-8829-03e86eb0c90c\") " pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.023243 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c8c765f3-89eb-4077-8829-03e86eb0c90c-mcd-auth-proxy-config\") pod \"machine-config-daemon-ppngk\" (UID: \"c8c765f3-89eb-4077-8829-03e86eb0c90c\") " pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.024096 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c8c765f3-89eb-4077-8829-03e86eb0c90c-mcd-auth-proxy-config\") pod \"machine-config-daemon-ppngk\" (UID: \"c8c765f3-89eb-4077-8829-03e86eb0c90c\") " pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.031663 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.050245 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.063809 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.066951 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbd9b\" (UniqueName: \"kubernetes.io/projected/f846b283-5468-4014-ba05-da5bfffa2ebc-kube-api-access-gbd9b\") pod \"node-resolver-46dqc\" (UID: \"f846b283-5468-4014-ba05-da5bfffa2ebc\") " pod="openshift-dns/node-resolver-46dqc" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.066902 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c8c765f3-89eb-4077-8829-03e86eb0c90c-proxy-tls\") pod \"machine-config-daemon-ppngk\" (UID: \"c8c765f3-89eb-4077-8829-03e86eb0c90c\") " pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.068401 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsqf5\" (UniqueName: \"kubernetes.io/projected/c8c765f3-89eb-4077-8829-03e86eb0c90c-kube-api-access-bsqf5\") pod \"machine-config-daemon-ppngk\" (UID: \"c8c765f3-89eb-4077-8829-03e86eb0c90c\") " pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.073553 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.087765 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.091748 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-46dqc" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.100160 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:10:16 crc kubenswrapper[4714]: W0129 16:10:16.101858 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf846b283_5468_4014_ba05_da5bfffa2ebc.slice/crio-d040290de4b9cff5da54cae43fa182eb438f5fb90c018e6aad54296b802f951c WatchSource:0}: Error finding container d040290de4b9cff5da54cae43fa182eb438f5fb90c018e6aad54296b802f951c: Status 404 returned error can't find the container with id d040290de4b9cff5da54cae43fa182eb438f5fb90c018e6aad54296b802f951c Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.125831 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:16 crc kubenswrapper[4714]: E0129 16:10:16.126073 4714 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 16:10:16 crc kubenswrapper[4714]: E0129 16:10:16.126167 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:18.126144867 +0000 UTC m=+24.646645987 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.126180 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.126310 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:16 crc kubenswrapper[4714]: E0129 16:10:16.126412 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.126428 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:16 crc kubenswrapper[4714]: E0129 16:10:16.126441 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 16:10:16 crc kubenswrapper[4714]: E0129 16:10:16.126518 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 16:10:16 crc kubenswrapper[4714]: E0129 16:10:16.126540 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 16:10:16 crc kubenswrapper[4714]: E0129 16:10:16.126541 4714 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:16 crc kubenswrapper[4714]: E0129 16:10:16.126554 4714 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:16 crc kubenswrapper[4714]: E0129 16:10:16.126606 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:18.12659419 +0000 UTC m=+24.647095310 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:16 crc kubenswrapper[4714]: E0129 16:10:16.126672 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:18.126640681 +0000 UTC m=+24.647141841 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:16 crc kubenswrapper[4714]: E0129 16:10:16.126678 4714 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 16:10:16 crc kubenswrapper[4714]: E0129 16:10:16.126757 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:18.126743134 +0000 UTC m=+24.647244294 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.135684 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 04:32:59.299196716 +0000 UTC Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.168096 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-2cfxk"] Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.169640 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.173024 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-b2ttm"] Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.173642 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.174824 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.175439 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.175880 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.176139 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.176327 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.176760 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.177375 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.194984 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.195959 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.196604 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.199145 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.202586 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.203697 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.205437 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.206198 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.207752 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.208543 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.209229 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.210497 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.211678 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.213918 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.215395 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.216332 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.217002 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.218154 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.220334 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.221323 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.222004 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.223503 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.224095 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.224771 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.225798 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.226599 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.227794 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.229076 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.231134 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.231349 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.231887 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.233961 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.234756 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.236092 4714 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.236733 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.239193 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.239928 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.240991 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.242572 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.243441 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.244090 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.244383 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.245444 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.246581 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.247123 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.248261 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.249041 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.250132 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.250673 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.251770 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.252423 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.254498 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.255516 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.255684 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.256406 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.258071 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.259185 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.260688 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.261373 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.267967 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.281553 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.296391 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.302290 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" event={"ID":"c8c765f3-89eb-4077-8829-03e86eb0c90c","Type":"ContainerStarted","Data":"27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5"} Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.302329 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" event={"ID":"c8c765f3-89eb-4077-8829-03e86eb0c90c","Type":"ContainerStarted","Data":"4b5e0d6b66078d9a0f57472436f75eb3078eba951da5242e7a34f9bb0dab7f27"} Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.302859 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-46dqc" event={"ID":"f846b283-5468-4014-ba05-da5bfffa2ebc","Type":"ContainerStarted","Data":"d040290de4b9cff5da54cae43fa182eb438f5fb90c018e6aad54296b802f951c"} Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.310469 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.325727 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.331360 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.331466 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-multus-socket-dir-parent\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.331517 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-cnibin\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.331553 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-hostroot\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.331977 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-cni-binary-copy\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.332136 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-cnibin\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.332267 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-host-run-netns\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.332421 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/89560008-8bdc-4640-af11-681d825e69d4-cni-binary-copy\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.332466 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-system-cni-dir\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.332521 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-multus-cni-dir\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.332589 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/89560008-8bdc-4640-af11-681d825e69d4-multus-daemon-config\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.332638 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkc7b\" (UniqueName: \"kubernetes.io/projected/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-kube-api-access-vkc7b\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.334438 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-os-release\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.334514 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-system-cni-dir\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.334543 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.334569 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-multus-conf-dir\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.334596 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-os-release\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.334622 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp6mh\" (UniqueName: \"kubernetes.io/projected/89560008-8bdc-4640-af11-681d825e69d4-kube-api-access-dp6mh\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.334647 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-host-run-k8s-cni-cncf-io\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.334672 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-host-var-lib-kubelet\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.334694 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-etc-kubernetes\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.334720 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-host-var-lib-cni-bin\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.334741 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-host-var-lib-cni-multus\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.334763 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-host-run-multus-certs\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.339674 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.353332 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.370319 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.389035 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.405866 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.420625 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.435329 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.435454 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.435494 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-multus-socket-dir-parent\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.435522 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-cnibin\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.435541 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-hostroot\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.435564 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-cni-binary-copy\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.435582 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-cnibin\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.435640 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-hostroot\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.435651 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-cnibin\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.435734 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-multus-socket-dir-parent\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.435763 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-cnibin\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.436235 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.436260 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-cni-binary-copy\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.436320 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-host-run-netns\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.436339 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/89560008-8bdc-4640-af11-681d825e69d4-cni-binary-copy\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.436394 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-host-run-netns\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.436356 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkc7b\" (UniqueName: \"kubernetes.io/projected/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-kube-api-access-vkc7b\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.436447 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-system-cni-dir\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.436716 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-system-cni-dir\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.436751 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-multus-cni-dir\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.436768 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/89560008-8bdc-4640-af11-681d825e69d4-multus-daemon-config\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.436831 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/89560008-8bdc-4640-af11-681d825e69d4-cni-binary-copy\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.436904 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-multus-cni-dir\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437090 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-os-release\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.436784 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-os-release\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437139 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-system-cni-dir\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437159 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437230 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-system-cni-dir\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437236 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/89560008-8bdc-4640-af11-681d825e69d4-multus-daemon-config\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437175 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-multus-conf-dir\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437291 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-os-release\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437307 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-host-run-k8s-cni-cncf-io\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437354 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-multus-conf-dir\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437397 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-os-release\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437429 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-host-var-lib-kubelet\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437446 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp6mh\" (UniqueName: \"kubernetes.io/projected/89560008-8bdc-4640-af11-681d825e69d4-kube-api-access-dp6mh\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437492 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-host-run-k8s-cni-cncf-io\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437522 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-host-var-lib-kubelet\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437546 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-host-var-lib-cni-bin\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437563 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-host-var-lib-cni-multus\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437614 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-host-var-lib-cni-bin\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437676 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-host-var-lib-cni-multus\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437579 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-host-run-multus-certs\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437797 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-host-run-multus-certs\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437802 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-etc-kubernetes\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437809 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.437847 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/89560008-8bdc-4640-af11-681d825e69d4-etc-kubernetes\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.447656 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.454881 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp6mh\" (UniqueName: \"kubernetes.io/projected/89560008-8bdc-4640-af11-681d825e69d4-kube-api-access-dp6mh\") pod \"multus-b2ttm\" (UID: \"89560008-8bdc-4640-af11-681d825e69d4\") " pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.456604 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkc7b\" (UniqueName: \"kubernetes.io/projected/8b20fd8d-1ebb-47d0-8676-403b99dac1ec-kube-api-access-vkc7b\") pod \"multus-additional-cni-plugins-2cfxk\" (UID: \"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\") " pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.462005 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.476351 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.479498 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.489806 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.494342 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.501321 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.503613 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.514504 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.523120 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.533886 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: W0129 16:10:16.535188 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b20fd8d_1ebb_47d0_8676_403b99dac1ec.slice/crio-62a298b4ef62dc42bfcda4ebb4fbb3d14f082ff2b065650707ddc28f29e2676c WatchSource:0}: Error finding container 62a298b4ef62dc42bfcda4ebb4fbb3d14f082ff2b065650707ddc28f29e2676c: Status 404 returned error can't find the container with id 62a298b4ef62dc42bfcda4ebb4fbb3d14f082ff2b065650707ddc28f29e2676c Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.539228 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-b2ttm" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.547237 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.549961 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-sbnkt"] Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.550721 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: W0129 16:10:16.554721 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89560008_8bdc_4640_af11_681d825e69d4.slice/crio-f7e9b13edf785c5d027ffdc99f2efd71174884823eda661e75fcd22c34590d7a WatchSource:0}: Error finding container f7e9b13edf785c5d027ffdc99f2efd71174884823eda661e75fcd22c34590d7a: Status 404 returned error can't find the container with id f7e9b13edf785c5d027ffdc99f2efd71174884823eda661e75fcd22c34590d7a Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.554854 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.555007 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.555771 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.556038 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.556378 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.556370 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.560152 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.564878 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.583603 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.609280 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.627748 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.640543 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-kubelet\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.640605 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-run-netns\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.640635 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.640661 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/04b20f02-6c1e-4082-8233-8f06bda63195-env-overrides\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.640683 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-cni-netd\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.640719 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-run-systemd\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.640743 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-run-ovn\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.640764 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-node-log\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.640781 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/04b20f02-6c1e-4082-8233-8f06bda63195-ovnkube-config\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.640804 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/04b20f02-6c1e-4082-8233-8f06bda63195-ovn-node-metrics-cert\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.640847 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-systemd-units\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.640867 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-cni-bin\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.640896 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-etc-openvswitch\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.640917 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-log-socket\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.640953 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/04b20f02-6c1e-4082-8233-8f06bda63195-ovnkube-script-lib\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.640970 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vrsm\" (UniqueName: \"kubernetes.io/projected/04b20f02-6c1e-4082-8233-8f06bda63195-kube-api-access-7vrsm\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.640989 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-var-lib-openvswitch\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.641012 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-slash\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.641040 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-run-ovn-kubernetes\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.641058 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-run-openvswitch\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.647837 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.681738 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.697002 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.711874 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.724918 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.735009 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.741772 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/04b20f02-6c1e-4082-8233-8f06bda63195-env-overrides\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.741820 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-cni-netd\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.741859 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-run-systemd\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.741885 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-run-ovn\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.741911 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/04b20f02-6c1e-4082-8233-8f06bda63195-ovn-node-metrics-cert\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.741983 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-node-log\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.741983 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-cni-netd\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742006 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/04b20f02-6c1e-4082-8233-8f06bda63195-ovnkube-config\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742026 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-run-systemd\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742067 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-systemd-units\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742061 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-run-ovn\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742105 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-systemd-units\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742116 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-node-log\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742183 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-cni-bin\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742216 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-etc-openvswitch\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742234 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-log-socket\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742252 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/04b20f02-6c1e-4082-8233-8f06bda63195-ovnkube-script-lib\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742268 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vrsm\" (UniqueName: \"kubernetes.io/projected/04b20f02-6c1e-4082-8233-8f06bda63195-kube-api-access-7vrsm\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742290 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-var-lib-openvswitch\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742323 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-slash\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742340 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-run-ovn-kubernetes\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742361 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-run-openvswitch\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742377 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-kubelet\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742394 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-run-netns\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742418 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742474 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742500 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-cni-bin\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742521 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-etc-openvswitch\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.742543 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-log-socket\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.743351 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-run-ovn-kubernetes\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.743411 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/04b20f02-6c1e-4082-8233-8f06bda63195-ovnkube-config\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.743491 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-slash\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.743501 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-kubelet\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.743517 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/04b20f02-6c1e-4082-8233-8f06bda63195-ovnkube-script-lib\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.743533 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-run-openvswitch\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.743511 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-var-lib-openvswitch\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.743554 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-run-netns\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.743627 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/04b20f02-6c1e-4082-8233-8f06bda63195-env-overrides\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.748557 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/04b20f02-6c1e-4082-8233-8f06bda63195-ovn-node-metrics-cert\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.758870 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.780988 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vrsm\" (UniqueName: \"kubernetes.io/projected/04b20f02-6c1e-4082-8233-8f06bda63195-kube-api-access-7vrsm\") pod \"ovnkube-node-sbnkt\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.819075 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.856756 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.874347 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:16 crc kubenswrapper[4714]: W0129 16:10:16.889140 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04b20f02_6c1e_4082_8233_8f06bda63195.slice/crio-17aad70fcdfcfc2aa07f37d1c4b0d894a800d6ca4c4b34e6100a73fad699fe31 WatchSource:0}: Error finding container 17aad70fcdfcfc2aa07f37d1c4b0d894a800d6ca4c4b34e6100a73fad699fe31: Status 404 returned error can't find the container with id 17aad70fcdfcfc2aa07f37d1c4b0d894a800d6ca4c4b34e6100a73fad699fe31 Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.909023 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.942720 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:16 crc kubenswrapper[4714]: I0129 16:10:16.978701 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.016900 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.060514 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.096299 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.136787 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 23:14:33.032377679 +0000 UTC Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.137830 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.178126 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.183250 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.183300 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.183310 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:17 crc kubenswrapper[4714]: E0129 16:10:17.183416 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:17 crc kubenswrapper[4714]: E0129 16:10:17.183572 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:17 crc kubenswrapper[4714]: E0129 16:10:17.183697 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.228753 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.260669 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.300811 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.309077 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-46dqc" event={"ID":"f846b283-5468-4014-ba05-da5bfffa2ebc","Type":"ContainerStarted","Data":"4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c"} Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.311332 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" event={"ID":"c8c765f3-89eb-4077-8829-03e86eb0c90c","Type":"ContainerStarted","Data":"bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54"} Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.312747 4714 generic.go:334] "Generic (PLEG): container finished" podID="04b20f02-6c1e-4082-8233-8f06bda63195" containerID="6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba" exitCode=0 Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.312836 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerDied","Data":"6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba"} Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.312895 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerStarted","Data":"17aad70fcdfcfc2aa07f37d1c4b0d894a800d6ca4c4b34e6100a73fad699fe31"} Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.314761 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b2ttm" event={"ID":"89560008-8bdc-4640-af11-681d825e69d4","Type":"ContainerStarted","Data":"a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a"} Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.314812 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b2ttm" event={"ID":"89560008-8bdc-4640-af11-681d825e69d4","Type":"ContainerStarted","Data":"f7e9b13edf785c5d027ffdc99f2efd71174884823eda661e75fcd22c34590d7a"} Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.317824 4714 generic.go:334] "Generic (PLEG): container finished" podID="8b20fd8d-1ebb-47d0-8676-403b99dac1ec" containerID="e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76" exitCode=0 Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.318816 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" event={"ID":"8b20fd8d-1ebb-47d0-8676-403b99dac1ec","Type":"ContainerDied","Data":"e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76"} Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.318870 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" event={"ID":"8b20fd8d-1ebb-47d0-8676-403b99dac1ec","Type":"ContainerStarted","Data":"62a298b4ef62dc42bfcda4ebb4fbb3d14f082ff2b065650707ddc28f29e2676c"} Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.339762 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.377570 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.420406 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.456837 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.496070 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.537972 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.577893 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.616862 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.664467 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.694204 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.738020 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.775807 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.820157 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.860331 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:17 crc kubenswrapper[4714]: I0129 16:10:17.909572 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.059995 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:10:18 crc kubenswrapper[4714]: E0129 16:10:18.060278 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:10:22.060252947 +0000 UTC m=+28.580754067 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.081392 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-c9jhc"] Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.081804 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-c9jhc" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.083430 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.083746 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.083997 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.084298 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.095986 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.113514 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.126483 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.137351 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 04:44:44.427951026 +0000 UTC Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.139894 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.160613 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.160660 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.160680 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.160705 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:18 crc kubenswrapper[4714]: E0129 16:10:18.160783 4714 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 16:10:18 crc kubenswrapper[4714]: E0129 16:10:18.160848 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 16:10:18 crc kubenswrapper[4714]: E0129 16:10:18.160866 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 16:10:18 crc kubenswrapper[4714]: E0129 16:10:18.160877 4714 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:18 crc kubenswrapper[4714]: E0129 16:10:18.160891 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:22.160863108 +0000 UTC m=+28.681364218 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 16:10:18 crc kubenswrapper[4714]: E0129 16:10:18.160941 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:22.160910799 +0000 UTC m=+28.681411919 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:18 crc kubenswrapper[4714]: E0129 16:10:18.160993 4714 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 16:10:18 crc kubenswrapper[4714]: E0129 16:10:18.161017 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:22.161011062 +0000 UTC m=+28.681512182 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 16:10:18 crc kubenswrapper[4714]: E0129 16:10:18.161049 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 16:10:18 crc kubenswrapper[4714]: E0129 16:10:18.161075 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 16:10:18 crc kubenswrapper[4714]: E0129 16:10:18.161094 4714 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:18 crc kubenswrapper[4714]: E0129 16:10:18.161146 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:22.161134056 +0000 UTC m=+28.681635176 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.186614 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.223414 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.256251 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.261628 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th2m6\" (UniqueName: \"kubernetes.io/projected/f80aba4c-9372-4bea-b537-cbd9b0a3e972-kube-api-access-th2m6\") pod \"node-ca-c9jhc\" (UID: \"f80aba4c-9372-4bea-b537-cbd9b0a3e972\") " pod="openshift-image-registry/node-ca-c9jhc" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.261700 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f80aba4c-9372-4bea-b537-cbd9b0a3e972-serviceca\") pod \"node-ca-c9jhc\" (UID: \"f80aba4c-9372-4bea-b537-cbd9b0a3e972\") " pod="openshift-image-registry/node-ca-c9jhc" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.261750 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f80aba4c-9372-4bea-b537-cbd9b0a3e972-host\") pod \"node-ca-c9jhc\" (UID: \"f80aba4c-9372-4bea-b537-cbd9b0a3e972\") " pod="openshift-image-registry/node-ca-c9jhc" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.295838 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.324496 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9"} Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.329013 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerStarted","Data":"e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4"} Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.329077 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerStarted","Data":"429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816"} Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.329093 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerStarted","Data":"b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc"} Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.329108 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerStarted","Data":"2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd"} Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.329119 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerStarted","Data":"7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016"} Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.331285 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" event={"ID":"8b20fd8d-1ebb-47d0-8676-403b99dac1ec","Type":"ContainerStarted","Data":"b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a"} Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.338691 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.362325 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f80aba4c-9372-4bea-b537-cbd9b0a3e972-serviceca\") pod \"node-ca-c9jhc\" (UID: \"f80aba4c-9372-4bea-b537-cbd9b0a3e972\") " pod="openshift-image-registry/node-ca-c9jhc" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.362404 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f80aba4c-9372-4bea-b537-cbd9b0a3e972-host\") pod \"node-ca-c9jhc\" (UID: \"f80aba4c-9372-4bea-b537-cbd9b0a3e972\") " pod="openshift-image-registry/node-ca-c9jhc" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.362436 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th2m6\" (UniqueName: \"kubernetes.io/projected/f80aba4c-9372-4bea-b537-cbd9b0a3e972-kube-api-access-th2m6\") pod \"node-ca-c9jhc\" (UID: \"f80aba4c-9372-4bea-b537-cbd9b0a3e972\") " pod="openshift-image-registry/node-ca-c9jhc" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.362588 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f80aba4c-9372-4bea-b537-cbd9b0a3e972-host\") pod \"node-ca-c9jhc\" (UID: \"f80aba4c-9372-4bea-b537-cbd9b0a3e972\") " pod="openshift-image-registry/node-ca-c9jhc" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.363761 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f80aba4c-9372-4bea-b537-cbd9b0a3e972-serviceca\") pod \"node-ca-c9jhc\" (UID: \"f80aba4c-9372-4bea-b537-cbd9b0a3e972\") " pod="openshift-image-registry/node-ca-c9jhc" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.379746 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.408878 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th2m6\" (UniqueName: \"kubernetes.io/projected/f80aba4c-9372-4bea-b537-cbd9b0a3e972-kube-api-access-th2m6\") pod \"node-ca-c9jhc\" (UID: \"f80aba4c-9372-4bea-b537-cbd9b0a3e972\") " pod="openshift-image-registry/node-ca-c9jhc" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.418419 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-c9jhc" Jan 29 16:10:18 crc kubenswrapper[4714]: W0129 16:10:18.434692 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf80aba4c_9372_4bea_b537_cbd9b0a3e972.slice/crio-e6137eb162566b71c4dcc93123d898651a75be1290e37dc810157e1ff49321df WatchSource:0}: Error finding container e6137eb162566b71c4dcc93123d898651a75be1290e37dc810157e1ff49321df: Status 404 returned error can't find the container with id e6137eb162566b71c4dcc93123d898651a75be1290e37dc810157e1ff49321df Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.443159 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.479059 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.517589 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.561766 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.595572 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.653075 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.683177 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.724249 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.758042 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.799772 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.838616 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.890605 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.923378 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.955690 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:18 crc kubenswrapper[4714]: I0129 16:10:18.998880 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:18Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.037780 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.083911 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.122910 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.137998 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 06:01:02.093902708 +0000 UTC Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.166078 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.183186 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.183239 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.183300 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:19 crc kubenswrapper[4714]: E0129 16:10:19.183323 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:19 crc kubenswrapper[4714]: E0129 16:10:19.183429 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:19 crc kubenswrapper[4714]: E0129 16:10:19.183688 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.209317 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.341208 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerStarted","Data":"5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7"} Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.342966 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-c9jhc" event={"ID":"f80aba4c-9372-4bea-b537-cbd9b0a3e972","Type":"ContainerStarted","Data":"cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e"} Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.343022 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-c9jhc" event={"ID":"f80aba4c-9372-4bea-b537-cbd9b0a3e972","Type":"ContainerStarted","Data":"e6137eb162566b71c4dcc93123d898651a75be1290e37dc810157e1ff49321df"} Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.345306 4714 generic.go:334] "Generic (PLEG): container finished" podID="8b20fd8d-1ebb-47d0-8676-403b99dac1ec" containerID="b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a" exitCode=0 Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.345384 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" event={"ID":"8b20fd8d-1ebb-47d0-8676-403b99dac1ec","Type":"ContainerDied","Data":"b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a"} Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.371480 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.393615 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.422775 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.441708 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.468977 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.481827 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.495212 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.514649 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.558066 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.599951 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.640799 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.682649 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.724621 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.763543 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.803454 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.844602 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.881149 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.920253 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:19 crc kubenswrapper[4714]: I0129 16:10:19.962816 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:19Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.008225 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.039270 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.079416 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.122324 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.138405 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 01:38:02.233237927 +0000 UTC Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.160839 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.204484 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.251064 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.281666 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.321482 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.357529 4714 generic.go:334] "Generic (PLEG): container finished" podID="8b20fd8d-1ebb-47d0-8676-403b99dac1ec" containerID="312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae" exitCode=0 Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.357658 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" event={"ID":"8b20fd8d-1ebb-47d0-8676-403b99dac1ec","Type":"ContainerDied","Data":"312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae"} Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.367736 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.402355 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.446140 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.479597 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.516441 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.560217 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.596101 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.639749 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.677196 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.722978 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.754596 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.777050 4714 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.778670 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.778706 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.778718 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.778834 4714 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.795524 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.850014 4714 kubelet_node_status.go:115] "Node was previously registered" node="crc" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.850497 4714 kubelet_node_status.go:79] "Successfully registered node" node="crc" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.852218 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.852281 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.852299 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.852324 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.852344 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:20Z","lastTransitionTime":"2026-01-29T16:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:20 crc kubenswrapper[4714]: E0129 16:10:20.871341 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.876472 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.876513 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.876531 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.876550 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.876562 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:20Z","lastTransitionTime":"2026-01-29T16:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.881323 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: E0129 16:10:20.891514 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.896301 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.896357 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.896376 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.896401 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.896419 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:20Z","lastTransitionTime":"2026-01-29T16:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:20 crc kubenswrapper[4714]: E0129 16:10:20.913419 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.916393 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.918115 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.918175 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.918191 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.918220 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.918237 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:20Z","lastTransitionTime":"2026-01-29T16:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:20 crc kubenswrapper[4714]: E0129 16:10:20.929441 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.934287 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.934350 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.934369 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.934398 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.934417 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:20Z","lastTransitionTime":"2026-01-29T16:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:20 crc kubenswrapper[4714]: E0129 16:10:20.952475 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:20 crc kubenswrapper[4714]: E0129 16:10:20.952664 4714 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.954811 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.954872 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.954894 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.954919 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.954959 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:20Z","lastTransitionTime":"2026-01-29T16:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:20 crc kubenswrapper[4714]: I0129 16:10:20.962778 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:20.999995 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:20Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.039600 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:21Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.057609 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.057676 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.057695 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.057721 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.057740 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:21Z","lastTransitionTime":"2026-01-29T16:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.139383 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 19:37:42.168647484 +0000 UTC Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.160554 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.160586 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.160599 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.160615 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.160627 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:21Z","lastTransitionTime":"2026-01-29T16:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.184072 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:21 crc kubenswrapper[4714]: E0129 16:10:21.184189 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.184553 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:21 crc kubenswrapper[4714]: E0129 16:10:21.184632 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.184706 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:21 crc kubenswrapper[4714]: E0129 16:10:21.184870 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.263425 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.263480 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.263499 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.263522 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.263539 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:21Z","lastTransitionTime":"2026-01-29T16:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.366123 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.366183 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.366206 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.366239 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.366320 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:21Z","lastTransitionTime":"2026-01-29T16:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.368430 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerStarted","Data":"662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf"} Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.372228 4714 generic.go:334] "Generic (PLEG): container finished" podID="8b20fd8d-1ebb-47d0-8676-403b99dac1ec" containerID="37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7" exitCode=0 Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.372278 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" event={"ID":"8b20fd8d-1ebb-47d0-8676-403b99dac1ec","Type":"ContainerDied","Data":"37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7"} Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.399803 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:21Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.421866 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:21Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.442573 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:21Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.462246 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:21Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.477175 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.477249 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.477276 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.477300 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.477317 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:21Z","lastTransitionTime":"2026-01-29T16:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.482219 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:21Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.500891 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:21Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.520620 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:21Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.540420 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:21Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.554452 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:21Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.569286 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:21Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.580272 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.580340 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.580367 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.580400 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.580425 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:21Z","lastTransitionTime":"2026-01-29T16:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.584117 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:21Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.635781 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:21Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.660501 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:21Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.683728 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.683760 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.683770 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.683783 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.683792 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:21Z","lastTransitionTime":"2026-01-29T16:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.684957 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:21Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.698300 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:21Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.786031 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.786070 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.786078 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.786093 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.786102 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:21Z","lastTransitionTime":"2026-01-29T16:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.888419 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.888460 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.888470 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.888485 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.888494 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:21Z","lastTransitionTime":"2026-01-29T16:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.992899 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.993001 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.993036 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.993086 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:21 crc kubenswrapper[4714]: I0129 16:10:21.993115 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:21Z","lastTransitionTime":"2026-01-29T16:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.096421 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.096488 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.096506 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.096534 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.096553 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:22Z","lastTransitionTime":"2026-01-29T16:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.111019 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:10:22 crc kubenswrapper[4714]: E0129 16:10:22.111288 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:10:30.111251322 +0000 UTC m=+36.631752482 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.140542 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 03:40:19.061605713 +0000 UTC Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.199758 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.199845 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.199870 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.199904 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.199928 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:22Z","lastTransitionTime":"2026-01-29T16:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.212235 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.212311 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.212382 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.212468 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:22 crc kubenswrapper[4714]: E0129 16:10:22.212552 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 16:10:22 crc kubenswrapper[4714]: E0129 16:10:22.212600 4714 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 16:10:22 crc kubenswrapper[4714]: E0129 16:10:22.212605 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 16:10:22 crc kubenswrapper[4714]: E0129 16:10:22.212642 4714 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:22 crc kubenswrapper[4714]: E0129 16:10:22.212683 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:30.212656906 +0000 UTC m=+36.733158066 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 16:10:22 crc kubenswrapper[4714]: E0129 16:10:22.212723 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:30.212696457 +0000 UTC m=+36.733197617 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:22 crc kubenswrapper[4714]: E0129 16:10:22.212812 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 16:10:22 crc kubenswrapper[4714]: E0129 16:10:22.212833 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 16:10:22 crc kubenswrapper[4714]: E0129 16:10:22.212849 4714 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:22 crc kubenswrapper[4714]: E0129 16:10:22.212893 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:30.212878683 +0000 UTC m=+36.733379843 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:22 crc kubenswrapper[4714]: E0129 16:10:22.213015 4714 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 16:10:22 crc kubenswrapper[4714]: E0129 16:10:22.213058 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:30.213044857 +0000 UTC m=+36.733546017 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.302739 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.302781 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.302796 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.302813 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.302825 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:22Z","lastTransitionTime":"2026-01-29T16:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.382067 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" event={"ID":"8b20fd8d-1ebb-47d0-8676-403b99dac1ec","Type":"ContainerStarted","Data":"93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59"} Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.405873 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.406001 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.406037 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.406068 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.406094 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:22Z","lastTransitionTime":"2026-01-29T16:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.409746 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:22Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.433721 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:22Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.455835 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:22Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.466998 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:22Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.487532 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:22Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.504609 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:22Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.508783 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.508812 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.508823 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.508838 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.508848 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:22Z","lastTransitionTime":"2026-01-29T16:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.521734 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:22Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.535006 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:22Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.560169 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:22Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.575010 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:22Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.590139 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:22Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.610461 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:22Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.611128 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.611196 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.611214 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.611238 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.611257 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:22Z","lastTransitionTime":"2026-01-29T16:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.623370 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:22Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.640218 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:22Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.666854 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:22Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.713918 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.714236 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.714248 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.714261 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.714270 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:22Z","lastTransitionTime":"2026-01-29T16:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.817105 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.817178 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.817198 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.817731 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.818020 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:22Z","lastTransitionTime":"2026-01-29T16:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.921014 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.921050 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.921068 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.921097 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:22 crc kubenswrapper[4714]: I0129 16:10:22.921117 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:22Z","lastTransitionTime":"2026-01-29T16:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.024730 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.024773 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.024784 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.024801 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.024816 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:23Z","lastTransitionTime":"2026-01-29T16:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.128235 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.128322 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.128348 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.128379 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.128401 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:23Z","lastTransitionTime":"2026-01-29T16:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.141612 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 11:36:46.142107128 +0000 UTC Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.184283 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.184320 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.184466 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:23 crc kubenswrapper[4714]: E0129 16:10:23.184458 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:23 crc kubenswrapper[4714]: E0129 16:10:23.184632 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:23 crc kubenswrapper[4714]: E0129 16:10:23.184784 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.233591 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.233654 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.233671 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.233691 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.233706 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:23Z","lastTransitionTime":"2026-01-29T16:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.337180 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.337238 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.337256 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.337280 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.337297 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:23Z","lastTransitionTime":"2026-01-29T16:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.390533 4714 generic.go:334] "Generic (PLEG): container finished" podID="8b20fd8d-1ebb-47d0-8676-403b99dac1ec" containerID="93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59" exitCode=0 Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.390632 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" event={"ID":"8b20fd8d-1ebb-47d0-8676-403b99dac1ec","Type":"ContainerDied","Data":"93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59"} Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.400495 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerStarted","Data":"7ce018ea1685c4d1a8769fd746ba32c24d3927e84ec15fd550a1a476e344ad5e"} Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.400980 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.401044 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.408060 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.425737 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.440736 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.440776 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.440784 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.440801 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.440811 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:23Z","lastTransitionTime":"2026-01-29T16:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.461226 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.466451 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.467291 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.488847 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.507716 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.525388 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.538673 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.543486 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.543510 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.543518 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.543533 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.543542 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:23Z","lastTransitionTime":"2026-01-29T16:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.549456 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.564305 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.575397 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.596526 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.612720 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.634870 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.647085 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.647120 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.647128 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.647143 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.647153 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:23Z","lastTransitionTime":"2026-01-29T16:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.648306 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.658642 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.671610 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.683372 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.695865 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.714636 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.727423 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.744098 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce018ea1685c4d1a8769fd746ba32c24d3927e84ec15fd550a1a476e344ad5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.749601 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.749629 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.749638 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.749662 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.749671 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:23Z","lastTransitionTime":"2026-01-29T16:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.754809 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.770534 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.781790 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.792192 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.802894 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.812593 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.823407 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.838906 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.852632 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.852678 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.852690 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.852707 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.852718 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:23Z","lastTransitionTime":"2026-01-29T16:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.859338 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:23Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.955270 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.955328 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.955347 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.955373 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.955394 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:23Z","lastTransitionTime":"2026-01-29T16:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:23 crc kubenswrapper[4714]: I0129 16:10:23.981492 4714 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.059033 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.059090 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.059107 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.059130 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.059146 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:24Z","lastTransitionTime":"2026-01-29T16:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.142596 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 20:25:44.147866268 +0000 UTC Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.162873 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.162957 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.162976 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.163001 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.163018 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:24Z","lastTransitionTime":"2026-01-29T16:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.201806 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.227687 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.245648 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.264622 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.264669 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.264682 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.264699 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.264712 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:24Z","lastTransitionTime":"2026-01-29T16:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.268700 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.287005 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.299966 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.331027 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce018ea1685c4d1a8769fd746ba32c24d3927e84ec15fd550a1a476e344ad5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.345365 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.360876 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.367131 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.367201 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.367221 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.367247 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.367266 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:24Z","lastTransitionTime":"2026-01-29T16:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.377695 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.388011 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.407224 4714 generic.go:334] "Generic (PLEG): container finished" podID="8b20fd8d-1ebb-47d0-8676-403b99dac1ec" containerID="1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2" exitCode=0 Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.407294 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" event={"ID":"8b20fd8d-1ebb-47d0-8676-403b99dac1ec","Type":"ContainerDied","Data":"1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2"} Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.407357 4714 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.410908 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.427779 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.442355 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.461773 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.470070 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.470120 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.470131 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.470150 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.470223 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:24Z","lastTransitionTime":"2026-01-29T16:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.484572 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.503015 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.515412 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.533675 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.557287 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.572869 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.572917 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.572951 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.572971 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.572983 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:24Z","lastTransitionTime":"2026-01-29T16:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.573271 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.592908 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.611354 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.626517 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.639351 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.670705 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce018ea1685c4d1a8769fd746ba32c24d3927e84ec15fd550a1a476e344ad5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.675501 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.675568 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.675594 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.675630 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.675656 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:24Z","lastTransitionTime":"2026-01-29T16:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.688349 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.704380 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.721675 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.736270 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.778104 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.778138 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.778150 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.778166 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.778178 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:24Z","lastTransitionTime":"2026-01-29T16:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.881605 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.881644 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.881655 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.881671 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.881683 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:24Z","lastTransitionTime":"2026-01-29T16:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.984703 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.984763 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.984780 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.984800 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:24 crc kubenswrapper[4714]: I0129 16:10:24.984817 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:24Z","lastTransitionTime":"2026-01-29T16:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.087996 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.088086 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.088106 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.088141 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.088166 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:25Z","lastTransitionTime":"2026-01-29T16:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.142987 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 03:54:54.259891104 +0000 UTC Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.183867 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.183909 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.183887 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:25 crc kubenswrapper[4714]: E0129 16:10:25.184130 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:25 crc kubenswrapper[4714]: E0129 16:10:25.184218 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:25 crc kubenswrapper[4714]: E0129 16:10:25.184321 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.190400 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.190429 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.190437 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.190449 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.190458 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:25Z","lastTransitionTime":"2026-01-29T16:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.292790 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.292828 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.292842 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.292861 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.292873 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:25Z","lastTransitionTime":"2026-01-29T16:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.395843 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.395882 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.395893 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.395909 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.395920 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:25Z","lastTransitionTime":"2026-01-29T16:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.413287 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" event={"ID":"8b20fd8d-1ebb-47d0-8676-403b99dac1ec","Type":"ContainerStarted","Data":"0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8"} Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.413314 4714 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.432363 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:25Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.446276 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:25Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.460896 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:25Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.478380 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:25Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.498559 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.498648 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.498675 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.498710 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.498738 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:25Z","lastTransitionTime":"2026-01-29T16:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.512086 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:25Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.530582 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:25Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.545869 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:25Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.565123 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:25Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.586717 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:25Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.602698 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.602777 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.602803 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.602834 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.602860 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:25Z","lastTransitionTime":"2026-01-29T16:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.605819 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:25Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.630126 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce018ea1685c4d1a8769fd746ba32c24d3927e84ec15fd550a1a476e344ad5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:25Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.646495 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:25Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.661790 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:25Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.681191 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:25Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.690853 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:25Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.706286 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.706505 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.706515 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.706530 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.706539 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:25Z","lastTransitionTime":"2026-01-29T16:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.808867 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.808947 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.808965 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.808986 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.809011 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:25Z","lastTransitionTime":"2026-01-29T16:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.912464 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.912862 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.913029 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.913137 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:25 crc kubenswrapper[4714]: I0129 16:10:25.913247 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:25Z","lastTransitionTime":"2026-01-29T16:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.021847 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.022166 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.022292 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.022385 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.022480 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:26Z","lastTransitionTime":"2026-01-29T16:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.125727 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.125783 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.125800 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.125821 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.125838 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:26Z","lastTransitionTime":"2026-01-29T16:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.144497 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 11:48:04.767504633 +0000 UTC Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.229016 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.229077 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.229095 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.229116 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.229133 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:26Z","lastTransitionTime":"2026-01-29T16:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.332612 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.332689 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.332716 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.332748 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.332771 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:26Z","lastTransitionTime":"2026-01-29T16:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.420447 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-sbnkt_04b20f02-6c1e-4082-8233-8f06bda63195/ovnkube-controller/0.log" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.424968 4714 generic.go:334] "Generic (PLEG): container finished" podID="04b20f02-6c1e-4082-8233-8f06bda63195" containerID="7ce018ea1685c4d1a8769fd746ba32c24d3927e84ec15fd550a1a476e344ad5e" exitCode=1 Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.425023 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerDied","Data":"7ce018ea1685c4d1a8769fd746ba32c24d3927e84ec15fd550a1a476e344ad5e"} Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.426194 4714 scope.go:117] "RemoveContainer" containerID="7ce018ea1685c4d1a8769fd746ba32c24d3927e84ec15fd550a1a476e344ad5e" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.434994 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.435047 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.435064 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.435087 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.435105 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:26Z","lastTransitionTime":"2026-01-29T16:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.456791 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:26Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.478446 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:26Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.495156 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:26Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.515709 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:26Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.535354 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:26Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.537746 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.537820 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.537843 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.537870 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.537888 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:26Z","lastTransitionTime":"2026-01-29T16:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.550633 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:26Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.571525 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce018ea1685c4d1a8769fd746ba32c24d3927e84ec15fd550a1a476e344ad5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ce018ea1685c4d1a8769fd746ba32c24d3927e84ec15fd550a1a476e344ad5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"message\\\":\\\"9 16:10:25.829056 6004 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0129 16:10:25.829066 6004 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0129 16:10:25.829080 6004 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 16:10:25.829136 6004 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0129 16:10:25.829145 6004 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0129 16:10:25.829167 6004 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0129 16:10:25.829172 6004 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0129 16:10:25.829187 6004 factory.go:656] Stopping watch factory\\\\nI0129 16:10:25.829200 6004 ovnkube.go:599] Stopped ovnkube\\\\nI0129 16:10:25.829231 6004 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 16:10:25.829240 6004 handler.go:208] Removed *v1.Node event handler 2\\\\nI0129 16:10:25.829249 6004 handler.go:208] Removed *v1.Node event handler 7\\\\nI0129 16:10:25.829257 6004 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0129 16:10:25.829264 6004 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0129 16:10:25.829277 6004 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0129 16:10:25.829285 6004 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0129 16\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:26Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.593046 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:26Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.620476 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:26Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.639112 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:26Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.640023 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.640057 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.640067 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.640082 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.640091 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:26Z","lastTransitionTime":"2026-01-29T16:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.654459 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:26Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.669745 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:26Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.691840 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:26Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.720001 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:26Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.742588 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.742644 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.742655 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.742673 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.742686 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:26Z","lastTransitionTime":"2026-01-29T16:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.745046 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:26Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.857526 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.858222 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.858279 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.858310 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.858331 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:26Z","lastTransitionTime":"2026-01-29T16:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.961611 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.961644 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.961654 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.961670 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:26 crc kubenswrapper[4714]: I0129 16:10:26.961681 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:26Z","lastTransitionTime":"2026-01-29T16:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.065431 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.065503 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.065538 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.065569 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.065592 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:27Z","lastTransitionTime":"2026-01-29T16:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.129718 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.145547 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 07:32:17.372594995 +0000 UTC Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.148649 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.163311 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.167648 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.167690 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.167703 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.167723 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.167755 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:27Z","lastTransitionTime":"2026-01-29T16:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.174534 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.183178 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.183233 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.183260 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:27 crc kubenswrapper[4714]: E0129 16:10:27.183585 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:27 crc kubenswrapper[4714]: E0129 16:10:27.183734 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:27 crc kubenswrapper[4714]: E0129 16:10:27.183922 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.188830 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.210360 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.226323 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.242150 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.255663 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.269586 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.271467 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.271526 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.271546 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.271570 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.271589 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:27Z","lastTransitionTime":"2026-01-29T16:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.285055 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.309133 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce018ea1685c4d1a8769fd746ba32c24d3927e84ec15fd550a1a476e344ad5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ce018ea1685c4d1a8769fd746ba32c24d3927e84ec15fd550a1a476e344ad5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"message\\\":\\\"9 16:10:25.829056 6004 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0129 16:10:25.829066 6004 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0129 16:10:25.829080 6004 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 16:10:25.829136 6004 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0129 16:10:25.829145 6004 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0129 16:10:25.829167 6004 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0129 16:10:25.829172 6004 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0129 16:10:25.829187 6004 factory.go:656] Stopping watch factory\\\\nI0129 16:10:25.829200 6004 ovnkube.go:599] Stopped ovnkube\\\\nI0129 16:10:25.829231 6004 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 16:10:25.829240 6004 handler.go:208] Removed *v1.Node event handler 2\\\\nI0129 16:10:25.829249 6004 handler.go:208] Removed *v1.Node event handler 7\\\\nI0129 16:10:25.829257 6004 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0129 16:10:25.829264 6004 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0129 16:10:25.829277 6004 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0129 16:10:25.829285 6004 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0129 16\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.324719 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.341037 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.367280 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.373341 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.373410 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.373431 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.373455 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.373471 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:27Z","lastTransitionTime":"2026-01-29T16:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.380539 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.431606 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-sbnkt_04b20f02-6c1e-4082-8233-8f06bda63195/ovnkube-controller/0.log" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.435915 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerStarted","Data":"5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075"} Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.436064 4714 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.469831 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.484448 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.484518 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.484535 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.484559 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.484581 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:27Z","lastTransitionTime":"2026-01-29T16:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.488031 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.500865 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.515700 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.528903 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.542867 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.559184 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.571236 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.587150 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.587197 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.587208 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.587225 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.587236 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:27Z","lastTransitionTime":"2026-01-29T16:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.591967 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ce018ea1685c4d1a8769fd746ba32c24d3927e84ec15fd550a1a476e344ad5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"message\\\":\\\"9 16:10:25.829056 6004 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0129 16:10:25.829066 6004 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0129 16:10:25.829080 6004 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 16:10:25.829136 6004 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0129 16:10:25.829145 6004 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0129 16:10:25.829167 6004 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0129 16:10:25.829172 6004 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0129 16:10:25.829187 6004 factory.go:656] Stopping watch factory\\\\nI0129 16:10:25.829200 6004 ovnkube.go:599] Stopped ovnkube\\\\nI0129 16:10:25.829231 6004 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 16:10:25.829240 6004 handler.go:208] Removed *v1.Node event handler 2\\\\nI0129 16:10:25.829249 6004 handler.go:208] Removed *v1.Node event handler 7\\\\nI0129 16:10:25.829257 6004 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0129 16:10:25.829264 6004 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0129 16:10:25.829277 6004 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0129 16:10:25.829285 6004 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0129 16\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.602897 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.612858 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.622029 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.632861 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.644160 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.657879 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:27Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.689689 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.689753 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.689767 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.689787 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.689802 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:27Z","lastTransitionTime":"2026-01-29T16:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.793163 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.793228 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.793245 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.793272 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.793292 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:27Z","lastTransitionTime":"2026-01-29T16:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.896744 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.896857 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.896881 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.896911 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:27 crc kubenswrapper[4714]: I0129 16:10:27.896962 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:27Z","lastTransitionTime":"2026-01-29T16:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.000527 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.000597 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.000627 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.000659 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.000682 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:28Z","lastTransitionTime":"2026-01-29T16:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.105057 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.105168 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.105187 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.105213 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.105234 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:28Z","lastTransitionTime":"2026-01-29T16:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.146455 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 15:24:40.250278632 +0000 UTC Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.207861 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.207904 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.207914 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.207949 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.207961 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:28Z","lastTransitionTime":"2026-01-29T16:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.310599 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.310645 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.310656 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.310673 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.310685 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:28Z","lastTransitionTime":"2026-01-29T16:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.413957 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.414007 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.414020 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.414044 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.414060 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:28Z","lastTransitionTime":"2026-01-29T16:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.442396 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-sbnkt_04b20f02-6c1e-4082-8233-8f06bda63195/ovnkube-controller/1.log" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.443410 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-sbnkt_04b20f02-6c1e-4082-8233-8f06bda63195/ovnkube-controller/0.log" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.447882 4714 generic.go:334] "Generic (PLEG): container finished" podID="04b20f02-6c1e-4082-8233-8f06bda63195" containerID="5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075" exitCode=1 Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.448009 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerDied","Data":"5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075"} Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.448116 4714 scope.go:117] "RemoveContainer" containerID="7ce018ea1685c4d1a8769fd746ba32c24d3927e84ec15fd550a1a476e344ad5e" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.448807 4714 scope.go:117] "RemoveContainer" containerID="5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075" Jan 29 16:10:28 crc kubenswrapper[4714]: E0129 16:10:28.449031 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-sbnkt_openshift-ovn-kubernetes(04b20f02-6c1e-4082-8233-8f06bda63195)\"" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.483328 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:28Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.505382 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:28Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.516926 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.517011 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.517029 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.517055 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.517078 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:28Z","lastTransitionTime":"2026-01-29T16:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.527708 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:28Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.543701 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:28Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.562670 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:28Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.579903 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:28Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.620519 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.620590 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.620615 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.620645 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.620668 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:28Z","lastTransitionTime":"2026-01-29T16:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.620831 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ce018ea1685c4d1a8769fd746ba32c24d3927e84ec15fd550a1a476e344ad5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"message\\\":\\\"9 16:10:25.829056 6004 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0129 16:10:25.829066 6004 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0129 16:10:25.829080 6004 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 16:10:25.829136 6004 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0129 16:10:25.829145 6004 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0129 16:10:25.829167 6004 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0129 16:10:25.829172 6004 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0129 16:10:25.829187 6004 factory.go:656] Stopping watch factory\\\\nI0129 16:10:25.829200 6004 ovnkube.go:599] Stopped ovnkube\\\\nI0129 16:10:25.829231 6004 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 16:10:25.829240 6004 handler.go:208] Removed *v1.Node event handler 2\\\\nI0129 16:10:25.829249 6004 handler.go:208] Removed *v1.Node event handler 7\\\\nI0129 16:10:25.829257 6004 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0129 16:10:25.829264 6004 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0129 16:10:25.829277 6004 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0129 16:10:25.829285 6004 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0129 16\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"message\\\":\\\"able:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 16:10:27.532704 6150 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-c9jhc\\\\nI0129 16:10:27.533048 6150 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0129 16:10:27.533063 6150 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:28Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.642650 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:28Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.662478 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:28Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.686790 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:28Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.699739 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:28Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.723078 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:28Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.723901 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.723991 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.724016 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.724050 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.724073 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:28Z","lastTransitionTime":"2026-01-29T16:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.745677 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:28Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.765119 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:28Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.779142 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:28Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.828473 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.828537 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.828556 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.828581 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.828599 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:28Z","lastTransitionTime":"2026-01-29T16:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.932248 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.932342 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.932363 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.932388 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:28 crc kubenswrapper[4714]: I0129 16:10:28.932439 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:28Z","lastTransitionTime":"2026-01-29T16:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.035138 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.035209 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.035229 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.035256 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.035279 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:29Z","lastTransitionTime":"2026-01-29T16:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.138039 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.138102 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.138120 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.138144 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.138161 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:29Z","lastTransitionTime":"2026-01-29T16:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.147445 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 08:14:55.878801094 +0000 UTC Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.184091 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.184168 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.184194 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:29 crc kubenswrapper[4714]: E0129 16:10:29.184269 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:29 crc kubenswrapper[4714]: E0129 16:10:29.184496 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:29 crc kubenswrapper[4714]: E0129 16:10:29.184770 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.241438 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.241514 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.241540 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.241571 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.241594 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:29Z","lastTransitionTime":"2026-01-29T16:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.344258 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.344325 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.344342 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.344364 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.344381 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:29Z","lastTransitionTime":"2026-01-29T16:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.424973 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw"] Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.425705 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.428271 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.428311 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.446909 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.446977 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.446990 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.447008 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.447018 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:29Z","lastTransitionTime":"2026-01-29T16:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.452814 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-sbnkt_04b20f02-6c1e-4082-8233-8f06bda63195/ovnkube-controller/1.log" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.455193 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.475497 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.487607 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2932c3bd-04c7-4494-8d43-03c4524a353f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tg8sw\" (UID: \"2932c3bd-04c7-4494-8d43-03c4524a353f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.487658 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2932c3bd-04c7-4494-8d43-03c4524a353f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tg8sw\" (UID: \"2932c3bd-04c7-4494-8d43-03c4524a353f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.487681 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvrtd\" (UniqueName: \"kubernetes.io/projected/2932c3bd-04c7-4494-8d43-03c4524a353f-kube-api-access-fvrtd\") pod \"ovnkube-control-plane-749d76644c-tg8sw\" (UID: \"2932c3bd-04c7-4494-8d43-03c4524a353f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.487707 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2932c3bd-04c7-4494-8d43-03c4524a353f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tg8sw\" (UID: \"2932c3bd-04c7-4494-8d43-03c4524a353f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.492367 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.510952 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.524584 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.535398 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.550236 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.550284 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.550297 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.550314 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.550328 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:29Z","lastTransitionTime":"2026-01-29T16:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.553166 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ce018ea1685c4d1a8769fd746ba32c24d3927e84ec15fd550a1a476e344ad5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"message\\\":\\\"9 16:10:25.829056 6004 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0129 16:10:25.829066 6004 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0129 16:10:25.829080 6004 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 16:10:25.829136 6004 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0129 16:10:25.829145 6004 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0129 16:10:25.829167 6004 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0129 16:10:25.829172 6004 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0129 16:10:25.829187 6004 factory.go:656] Stopping watch factory\\\\nI0129 16:10:25.829200 6004 ovnkube.go:599] Stopped ovnkube\\\\nI0129 16:10:25.829231 6004 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 16:10:25.829240 6004 handler.go:208] Removed *v1.Node event handler 2\\\\nI0129 16:10:25.829249 6004 handler.go:208] Removed *v1.Node event handler 7\\\\nI0129 16:10:25.829257 6004 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0129 16:10:25.829264 6004 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0129 16:10:25.829277 6004 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0129 16:10:25.829285 6004 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0129 16\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"message\\\":\\\"able:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 16:10:27.532704 6150 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-c9jhc\\\\nI0129 16:10:27.533048 6150 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0129 16:10:27.533063 6150 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.565439 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2932c3bd-04c7-4494-8d43-03c4524a353f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tg8sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.578877 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.588901 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2932c3bd-04c7-4494-8d43-03c4524a353f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tg8sw\" (UID: \"2932c3bd-04c7-4494-8d43-03c4524a353f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.589046 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2932c3bd-04c7-4494-8d43-03c4524a353f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tg8sw\" (UID: \"2932c3bd-04c7-4494-8d43-03c4524a353f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.589087 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2932c3bd-04c7-4494-8d43-03c4524a353f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tg8sw\" (UID: \"2932c3bd-04c7-4494-8d43-03c4524a353f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.589120 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvrtd\" (UniqueName: \"kubernetes.io/projected/2932c3bd-04c7-4494-8d43-03c4524a353f-kube-api-access-fvrtd\") pod \"ovnkube-control-plane-749d76644c-tg8sw\" (UID: \"2932c3bd-04c7-4494-8d43-03c4524a353f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.590210 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2932c3bd-04c7-4494-8d43-03c4524a353f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tg8sw\" (UID: \"2932c3bd-04c7-4494-8d43-03c4524a353f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.590785 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2932c3bd-04c7-4494-8d43-03c4524a353f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tg8sw\" (UID: \"2932c3bd-04c7-4494-8d43-03c4524a353f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.594102 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.596268 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2932c3bd-04c7-4494-8d43-03c4524a353f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tg8sw\" (UID: \"2932c3bd-04c7-4494-8d43-03c4524a353f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.613863 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.619797 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvrtd\" (UniqueName: \"kubernetes.io/projected/2932c3bd-04c7-4494-8d43-03c4524a353f-kube-api-access-fvrtd\") pod \"ovnkube-control-plane-749d76644c-tg8sw\" (UID: \"2932c3bd-04c7-4494-8d43-03c4524a353f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.627405 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.641982 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.652536 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.652580 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.652613 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.652631 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.652642 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:29Z","lastTransitionTime":"2026-01-29T16:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.660413 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.678154 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.696752 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.746697 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.755699 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.755765 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.755785 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.755811 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.755830 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:29Z","lastTransitionTime":"2026-01-29T16:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:29 crc kubenswrapper[4714]: W0129 16:10:29.767948 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2932c3bd_04c7_4494_8d43_03c4524a353f.slice/crio-e67687fd660061194ff48003829241962cc072844249c60589fca5b2a4a7d8ef WatchSource:0}: Error finding container e67687fd660061194ff48003829241962cc072844249c60589fca5b2a4a7d8ef: Status 404 returned error can't find the container with id e67687fd660061194ff48003829241962cc072844249c60589fca5b2a4a7d8ef Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.859437 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.859481 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.859493 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.859513 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.859526 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:29Z","lastTransitionTime":"2026-01-29T16:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.962318 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.962366 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.962383 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.962405 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:29 crc kubenswrapper[4714]: I0129 16:10:29.962421 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:29Z","lastTransitionTime":"2026-01-29T16:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.065651 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.065688 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.065701 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.065719 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.065732 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:30Z","lastTransitionTime":"2026-01-29T16:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.148341 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 10:07:15.895851809 +0000 UTC Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.168739 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-2w92b"] Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.169584 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.169604 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:30 crc kubenswrapper[4714]: E0129 16:10:30.169694 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.169720 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.169738 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.169757 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.169767 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:30Z","lastTransitionTime":"2026-01-29T16:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.184293 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.196115 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:10:30 crc kubenswrapper[4714]: E0129 16:10:30.196308 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:10:46.196276266 +0000 UTC m=+52.716777386 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.196451 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwv7p\" (UniqueName: \"kubernetes.io/projected/791456e8-8d95-4cdb-8fd1-d06a7586b328-kube-api-access-qwv7p\") pod \"network-metrics-daemon-2w92b\" (UID: \"791456e8-8d95-4cdb-8fd1-d06a7586b328\") " pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.196495 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs\") pod \"network-metrics-daemon-2w92b\" (UID: \"791456e8-8d95-4cdb-8fd1-d06a7586b328\") " pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.201522 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.221685 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.248478 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.259178 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.271887 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.271946 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.271957 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.271995 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.272007 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:30Z","lastTransitionTime":"2026-01-29T16:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.275696 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ce018ea1685c4d1a8769fd746ba32c24d3927e84ec15fd550a1a476e344ad5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"message\\\":\\\"9 16:10:25.829056 6004 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0129 16:10:25.829066 6004 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0129 16:10:25.829080 6004 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 16:10:25.829136 6004 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0129 16:10:25.829145 6004 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0129 16:10:25.829167 6004 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0129 16:10:25.829172 6004 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0129 16:10:25.829187 6004 factory.go:656] Stopping watch factory\\\\nI0129 16:10:25.829200 6004 ovnkube.go:599] Stopped ovnkube\\\\nI0129 16:10:25.829231 6004 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 16:10:25.829240 6004 handler.go:208] Removed *v1.Node event handler 2\\\\nI0129 16:10:25.829249 6004 handler.go:208] Removed *v1.Node event handler 7\\\\nI0129 16:10:25.829257 6004 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0129 16:10:25.829264 6004 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0129 16:10:25.829277 6004 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0129 16:10:25.829285 6004 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0129 16\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"message\\\":\\\"able:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 16:10:27.532704 6150 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-c9jhc\\\\nI0129 16:10:27.533048 6150 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0129 16:10:27.533063 6150 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.285284 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2932c3bd-04c7-4494-8d43-03c4524a353f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tg8sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.294887 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.297570 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.297622 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.297665 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.297696 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwv7p\" (UniqueName: \"kubernetes.io/projected/791456e8-8d95-4cdb-8fd1-d06a7586b328-kube-api-access-qwv7p\") pod \"network-metrics-daemon-2w92b\" (UID: \"791456e8-8d95-4cdb-8fd1-d06a7586b328\") " pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:30 crc kubenswrapper[4714]: E0129 16:10:30.297704 4714 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.297726 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs\") pod \"network-metrics-daemon-2w92b\" (UID: \"791456e8-8d95-4cdb-8fd1-d06a7586b328\") " pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:30 crc kubenswrapper[4714]: E0129 16:10:30.297762 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:46.297742792 +0000 UTC m=+52.818243912 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.297792 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:30 crc kubenswrapper[4714]: E0129 16:10:30.297826 4714 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 16:10:30 crc kubenswrapper[4714]: E0129 16:10:30.297874 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 16:10:30 crc kubenswrapper[4714]: E0129 16:10:30.297882 4714 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 16:10:30 crc kubenswrapper[4714]: E0129 16:10:30.297895 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:46.297878056 +0000 UTC m=+52.818379176 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 16:10:30 crc kubenswrapper[4714]: E0129 16:10:30.297901 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 16:10:30 crc kubenswrapper[4714]: E0129 16:10:30.297919 4714 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:30 crc kubenswrapper[4714]: E0129 16:10:30.297951 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 16:10:30 crc kubenswrapper[4714]: E0129 16:10:30.297965 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 16:10:30 crc kubenswrapper[4714]: E0129 16:10:30.297976 4714 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:30 crc kubenswrapper[4714]: E0129 16:10:30.297982 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs podName:791456e8-8d95-4cdb-8fd1-d06a7586b328 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:30.797959588 +0000 UTC m=+37.318460718 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs") pod "network-metrics-daemon-2w92b" (UID: "791456e8-8d95-4cdb-8fd1-d06a7586b328") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 16:10:30 crc kubenswrapper[4714]: E0129 16:10:30.298001 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:46.297994069 +0000 UTC m=+52.818495189 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:30 crc kubenswrapper[4714]: E0129 16:10:30.298021 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:46.29801311 +0000 UTC m=+52.818514230 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.314443 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwv7p\" (UniqueName: \"kubernetes.io/projected/791456e8-8d95-4cdb-8fd1-d06a7586b328-kube-api-access-qwv7p\") pod \"network-metrics-daemon-2w92b\" (UID: \"791456e8-8d95-4cdb-8fd1-d06a7586b328\") " pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.319829 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.332402 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.348061 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2w92b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791456e8-8d95-4cdb-8fd1-d06a7586b328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2w92b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.369758 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.375571 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.375597 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.375606 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.375621 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.375630 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:30Z","lastTransitionTime":"2026-01-29T16:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.393179 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.405998 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.419066 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.434538 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.448686 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.463817 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" event={"ID":"2932c3bd-04c7-4494-8d43-03c4524a353f","Type":"ContainerStarted","Data":"dc1d7b5f32fe8de4bfd5ba555838989293be538217424c86ea5cedadb8295f43"} Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.463875 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" event={"ID":"2932c3bd-04c7-4494-8d43-03c4524a353f","Type":"ContainerStarted","Data":"a78109e9e498f3d640934a2c45faf27133d32ece9e35e42ed48ddba720fa7a10"} Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.463887 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" event={"ID":"2932c3bd-04c7-4494-8d43-03c4524a353f","Type":"ContainerStarted","Data":"e67687fd660061194ff48003829241962cc072844249c60589fca5b2a4a7d8ef"} Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.479035 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.479081 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.479091 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.479109 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.479122 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:30Z","lastTransitionTime":"2026-01-29T16:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.482648 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.494833 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.508558 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.520708 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.529844 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2w92b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791456e8-8d95-4cdb-8fd1-d06a7586b328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2w92b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.542245 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.558832 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.571646 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.581246 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.581320 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.581337 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.581361 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.581375 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:30Z","lastTransitionTime":"2026-01-29T16:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.589382 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.617891 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.635167 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.658018 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.678492 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.684362 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.684421 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.684433 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.684464 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.684482 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:30Z","lastTransitionTime":"2026-01-29T16:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.699870 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.717875 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.742969 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ce018ea1685c4d1a8769fd746ba32c24d3927e84ec15fd550a1a476e344ad5e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"message\\\":\\\"9 16:10:25.829056 6004 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0129 16:10:25.829066 6004 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0129 16:10:25.829080 6004 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 16:10:25.829136 6004 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0129 16:10:25.829145 6004 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0129 16:10:25.829167 6004 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0129 16:10:25.829172 6004 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0129 16:10:25.829187 6004 factory.go:656] Stopping watch factory\\\\nI0129 16:10:25.829200 6004 ovnkube.go:599] Stopped ovnkube\\\\nI0129 16:10:25.829231 6004 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 16:10:25.829240 6004 handler.go:208] Removed *v1.Node event handler 2\\\\nI0129 16:10:25.829249 6004 handler.go:208] Removed *v1.Node event handler 7\\\\nI0129 16:10:25.829257 6004 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0129 16:10:25.829264 6004 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0129 16:10:25.829277 6004 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0129 16:10:25.829285 6004 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0129 16\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"message\\\":\\\"able:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 16:10:27.532704 6150 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-c9jhc\\\\nI0129 16:10:27.533048 6150 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0129 16:10:27.533063 6150 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.761478 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2932c3bd-04c7-4494-8d43-03c4524a353f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78109e9e498f3d640934a2c45faf27133d32ece9e35e42ed48ddba720fa7a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc1d7b5f32fe8de4bfd5ba555838989293be538217424c86ea5cedadb8295f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tg8sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:30Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.788270 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.788345 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.788373 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.788406 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.788432 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:30Z","lastTransitionTime":"2026-01-29T16:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.802347 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs\") pod \"network-metrics-daemon-2w92b\" (UID: \"791456e8-8d95-4cdb-8fd1-d06a7586b328\") " pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:30 crc kubenswrapper[4714]: E0129 16:10:30.802584 4714 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 16:10:30 crc kubenswrapper[4714]: E0129 16:10:30.802685 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs podName:791456e8-8d95-4cdb-8fd1-d06a7586b328 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:31.80265835 +0000 UTC m=+38.323159500 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs") pod "network-metrics-daemon-2w92b" (UID: "791456e8-8d95-4cdb-8fd1-d06a7586b328") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.892738 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.892789 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.892882 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.892919 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.892973 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:30Z","lastTransitionTime":"2026-01-29T16:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.979820 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.979907 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.979967 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.980003 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:30 crc kubenswrapper[4714]: I0129 16:10:30.980029 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:30Z","lastTransitionTime":"2026-01-29T16:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:31 crc kubenswrapper[4714]: E0129 16:10:31.008667 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:31Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.014844 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.014917 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.014970 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.015000 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.015027 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:31Z","lastTransitionTime":"2026-01-29T16:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:31 crc kubenswrapper[4714]: E0129 16:10:31.036590 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:31Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.042276 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.042348 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.042366 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.042396 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.042453 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:31Z","lastTransitionTime":"2026-01-29T16:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:31 crc kubenswrapper[4714]: E0129 16:10:31.063705 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:31Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.069475 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.069544 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.069562 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.069594 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.069613 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:31Z","lastTransitionTime":"2026-01-29T16:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:31 crc kubenswrapper[4714]: E0129 16:10:31.094569 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:31Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.100895 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.101004 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.101026 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.101061 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.101082 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:31Z","lastTransitionTime":"2026-01-29T16:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:31 crc kubenswrapper[4714]: E0129 16:10:31.124198 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:31Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:31 crc kubenswrapper[4714]: E0129 16:10:31.124466 4714 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.127522 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.127599 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.127624 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.127658 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.127684 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:31Z","lastTransitionTime":"2026-01-29T16:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.149182 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 15:28:44.721858485 +0000 UTC Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.183703 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.183703 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.183721 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:31 crc kubenswrapper[4714]: E0129 16:10:31.184045 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:31 crc kubenswrapper[4714]: E0129 16:10:31.184175 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:31 crc kubenswrapper[4714]: E0129 16:10:31.184395 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.230834 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.230915 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.230982 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.231020 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.231045 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:31Z","lastTransitionTime":"2026-01-29T16:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.334008 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.334071 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.334088 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.334111 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.334127 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:31Z","lastTransitionTime":"2026-01-29T16:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.437047 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.437309 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.437390 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.437483 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.437578 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:31Z","lastTransitionTime":"2026-01-29T16:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.540833 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.541164 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.541269 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.541366 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.541447 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:31Z","lastTransitionTime":"2026-01-29T16:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.644382 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.644793 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.645079 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.645406 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.645613 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:31Z","lastTransitionTime":"2026-01-29T16:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.748863 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.748926 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.748990 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.749032 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.749058 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:31Z","lastTransitionTime":"2026-01-29T16:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.814514 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs\") pod \"network-metrics-daemon-2w92b\" (UID: \"791456e8-8d95-4cdb-8fd1-d06a7586b328\") " pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:31 crc kubenswrapper[4714]: E0129 16:10:31.814741 4714 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 16:10:31 crc kubenswrapper[4714]: E0129 16:10:31.814868 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs podName:791456e8-8d95-4cdb-8fd1-d06a7586b328 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:33.814838924 +0000 UTC m=+40.335340084 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs") pod "network-metrics-daemon-2w92b" (UID: "791456e8-8d95-4cdb-8fd1-d06a7586b328") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.852439 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.852490 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.852508 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.852531 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.852548 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:31Z","lastTransitionTime":"2026-01-29T16:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.955187 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.955243 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.955261 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.955284 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:31 crc kubenswrapper[4714]: I0129 16:10:31.955303 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:31Z","lastTransitionTime":"2026-01-29T16:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.057654 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.057726 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.057754 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.057784 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.057808 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:32Z","lastTransitionTime":"2026-01-29T16:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.150100 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 22:40:02.755402892 +0000 UTC Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.160620 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.160688 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.160714 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.160740 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.160821 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:32Z","lastTransitionTime":"2026-01-29T16:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.183330 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:32 crc kubenswrapper[4714]: E0129 16:10:32.183596 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.250431 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.251377 4714 scope.go:117] "RemoveContainer" containerID="5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075" Jan 29 16:10:32 crc kubenswrapper[4714]: E0129 16:10:32.251720 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-sbnkt_openshift-ovn-kubernetes(04b20f02-6c1e-4082-8233-8f06bda63195)\"" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.264054 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.264111 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.264131 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.264154 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.264170 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:32Z","lastTransitionTime":"2026-01-29T16:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.278890 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:32Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.299147 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:32Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.313294 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:32Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.339200 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:32Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.365019 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:32Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.366503 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.366536 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.366549 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.366567 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.366580 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:32Z","lastTransitionTime":"2026-01-29T16:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.395668 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:32Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.418069 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"message\\\":\\\"able:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 16:10:27.532704 6150 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-c9jhc\\\\nI0129 16:10:27.533048 6150 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0129 16:10:27.533063 6150 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-sbnkt_openshift-ovn-kubernetes(04b20f02-6c1e-4082-8233-8f06bda63195)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:32Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.431499 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2932c3bd-04c7-4494-8d43-03c4524a353f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78109e9e498f3d640934a2c45faf27133d32ece9e35e42ed48ddba720fa7a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc1d7b5f32fe8de4bfd5ba555838989293be538217424c86ea5cedadb8295f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tg8sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:32Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.449141 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:32Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.469649 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.469689 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.469699 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.469715 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.469726 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:32Z","lastTransitionTime":"2026-01-29T16:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.473204 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:32Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.497684 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:32Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.515775 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:32Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.532373 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2w92b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791456e8-8d95-4cdb-8fd1-d06a7586b328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2w92b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:32Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.552563 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:32Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.572654 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.572815 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.572839 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.572880 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.572904 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:32Z","lastTransitionTime":"2026-01-29T16:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.573741 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:32Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.590780 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:32Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.610324 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:32Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.675759 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.675817 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.675835 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.675857 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.675874 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:32Z","lastTransitionTime":"2026-01-29T16:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.779516 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.779583 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.779602 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.779627 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.779644 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:32Z","lastTransitionTime":"2026-01-29T16:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.882074 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.882114 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.882125 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.882144 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.882156 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:32Z","lastTransitionTime":"2026-01-29T16:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.985449 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.985515 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.985533 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.985558 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:32 crc kubenswrapper[4714]: I0129 16:10:32.985575 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:32Z","lastTransitionTime":"2026-01-29T16:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.089042 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.089106 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.089129 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.089160 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.089183 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:33Z","lastTransitionTime":"2026-01-29T16:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.151106 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 22:54:09.081811871 +0000 UTC Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.183647 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.183676 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:33 crc kubenswrapper[4714]: E0129 16:10:33.183846 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.183682 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:33 crc kubenswrapper[4714]: E0129 16:10:33.184105 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:33 crc kubenswrapper[4714]: E0129 16:10:33.184218 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.192610 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.192716 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.192743 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.192775 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.192797 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:33Z","lastTransitionTime":"2026-01-29T16:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.296827 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.296874 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.296890 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.296911 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.296926 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:33Z","lastTransitionTime":"2026-01-29T16:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.399582 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.399643 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.399660 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.399687 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.399703 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:33Z","lastTransitionTime":"2026-01-29T16:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.503380 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.503435 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.503449 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.503469 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.503482 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:33Z","lastTransitionTime":"2026-01-29T16:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.606208 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.606250 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.606260 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.606275 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.606286 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:33Z","lastTransitionTime":"2026-01-29T16:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.708843 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.708972 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.708994 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.709018 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.709039 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:33Z","lastTransitionTime":"2026-01-29T16:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.812495 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.812566 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.812586 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.812612 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.812630 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:33Z","lastTransitionTime":"2026-01-29T16:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.835419 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs\") pod \"network-metrics-daemon-2w92b\" (UID: \"791456e8-8d95-4cdb-8fd1-d06a7586b328\") " pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:33 crc kubenswrapper[4714]: E0129 16:10:33.835609 4714 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 16:10:33 crc kubenswrapper[4714]: E0129 16:10:33.835691 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs podName:791456e8-8d95-4cdb-8fd1-d06a7586b328 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:37.835667981 +0000 UTC m=+44.356169131 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs") pod "network-metrics-daemon-2w92b" (UID: "791456e8-8d95-4cdb-8fd1-d06a7586b328") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.915412 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.915487 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.915506 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.915531 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:33 crc kubenswrapper[4714]: I0129 16:10:33.915548 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:33Z","lastTransitionTime":"2026-01-29T16:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.018704 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.018769 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.018789 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.018812 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.018848 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:34Z","lastTransitionTime":"2026-01-29T16:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.122819 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.122883 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.122901 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.122924 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.123037 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:34Z","lastTransitionTime":"2026-01-29T16:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.151573 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 21:36:20.812027632 +0000 UTC Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.184231 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:34 crc kubenswrapper[4714]: E0129 16:10:34.184825 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.218200 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"message\\\":\\\"able:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 16:10:27.532704 6150 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-c9jhc\\\\nI0129 16:10:27.533048 6150 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0129 16:10:27.533063 6150 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-sbnkt_openshift-ovn-kubernetes(04b20f02-6c1e-4082-8233-8f06bda63195)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:34Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.225631 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.225690 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.225709 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.225739 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.225757 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:34Z","lastTransitionTime":"2026-01-29T16:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.239198 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2932c3bd-04c7-4494-8d43-03c4524a353f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78109e9e498f3d640934a2c45faf27133d32ece9e35e42ed48ddba720fa7a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc1d7b5f32fe8de4bfd5ba555838989293be538217424c86ea5cedadb8295f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tg8sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:34Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.259188 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:34Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.280210 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:34Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.297806 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:34Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.313242 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2w92b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791456e8-8d95-4cdb-8fd1-d06a7586b328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2w92b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:34Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.330234 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.330344 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.330420 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.330507 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.330534 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:34Z","lastTransitionTime":"2026-01-29T16:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.331597 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:34Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.346269 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:34Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.370191 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:34Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.390385 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:34Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.406594 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:34Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.420217 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:34Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.430774 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:34Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.432872 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.432919 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.432957 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.432981 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.432993 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:34Z","lastTransitionTime":"2026-01-29T16:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.446994 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:34Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.460418 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:34Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.482000 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:34Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.493964 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:34Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.535027 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.535079 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.535088 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.535101 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.535111 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:34Z","lastTransitionTime":"2026-01-29T16:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.637508 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.637572 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.637589 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.637615 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.637635 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:34Z","lastTransitionTime":"2026-01-29T16:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.741241 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.741309 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.741335 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.741366 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.741388 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:34Z","lastTransitionTime":"2026-01-29T16:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.844546 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.844602 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.844621 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.844653 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.844676 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:34Z","lastTransitionTime":"2026-01-29T16:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.947746 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.947810 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.947826 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.947849 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:34 crc kubenswrapper[4714]: I0129 16:10:34.947866 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:34Z","lastTransitionTime":"2026-01-29T16:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.051110 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.051172 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.051190 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.051212 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.051228 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:35Z","lastTransitionTime":"2026-01-29T16:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.151730 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 22:52:08.830148651 +0000 UTC Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.153770 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.153817 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.153834 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.153863 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.153884 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:35Z","lastTransitionTime":"2026-01-29T16:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.183757 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.183846 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.183882 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:35 crc kubenswrapper[4714]: E0129 16:10:35.184003 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:35 crc kubenswrapper[4714]: E0129 16:10:35.184120 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:35 crc kubenswrapper[4714]: E0129 16:10:35.184224 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.255970 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.256003 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.256014 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.256027 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.256037 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:35Z","lastTransitionTime":"2026-01-29T16:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.358607 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.358686 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.358704 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.358728 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.358744 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:35Z","lastTransitionTime":"2026-01-29T16:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.461646 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.461714 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.461730 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.461756 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.461774 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:35Z","lastTransitionTime":"2026-01-29T16:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.565101 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.565210 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.565237 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.565309 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.565331 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:35Z","lastTransitionTime":"2026-01-29T16:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.668418 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.668489 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.668511 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.668546 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.668588 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:35Z","lastTransitionTime":"2026-01-29T16:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.771783 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.771840 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.771859 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.771889 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.771908 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:35Z","lastTransitionTime":"2026-01-29T16:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.875911 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.876007 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.876026 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.876050 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.876067 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:35Z","lastTransitionTime":"2026-01-29T16:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.979669 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.979736 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.979752 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.979776 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:35 crc kubenswrapper[4714]: I0129 16:10:35.979793 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:35Z","lastTransitionTime":"2026-01-29T16:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.082276 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.082351 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.082376 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.082410 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.082434 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:36Z","lastTransitionTime":"2026-01-29T16:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.152292 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 09:56:56.443823918 +0000 UTC Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.183197 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:36 crc kubenswrapper[4714]: E0129 16:10:36.183448 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.185503 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.185565 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.185581 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.185609 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.185627 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:36Z","lastTransitionTime":"2026-01-29T16:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.289316 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.289382 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.289398 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.289425 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.289442 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:36Z","lastTransitionTime":"2026-01-29T16:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.394611 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.394696 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.394723 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.394754 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.394779 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:36Z","lastTransitionTime":"2026-01-29T16:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.501692 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.501766 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.501785 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.501809 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.501827 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:36Z","lastTransitionTime":"2026-01-29T16:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.605129 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.605198 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.605219 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.605243 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.605261 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:36Z","lastTransitionTime":"2026-01-29T16:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.708697 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.708749 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.708762 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.708779 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.708790 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:36Z","lastTransitionTime":"2026-01-29T16:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.811866 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.811974 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.812003 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.812035 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.812056 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:36Z","lastTransitionTime":"2026-01-29T16:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.915895 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.916260 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.916276 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.916299 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:36 crc kubenswrapper[4714]: I0129 16:10:36.916326 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:36Z","lastTransitionTime":"2026-01-29T16:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.019578 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.019651 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.019673 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.019702 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.019719 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:37Z","lastTransitionTime":"2026-01-29T16:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.123230 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.123289 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.123305 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.123326 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.123343 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:37Z","lastTransitionTime":"2026-01-29T16:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.152713 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 20:11:28.548799934 +0000 UTC Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.183241 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.183274 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.183274 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:37 crc kubenswrapper[4714]: E0129 16:10:37.183453 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:37 crc kubenswrapper[4714]: E0129 16:10:37.183584 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:37 crc kubenswrapper[4714]: E0129 16:10:37.183669 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.226172 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.226257 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.226281 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.226311 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.226332 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:37Z","lastTransitionTime":"2026-01-29T16:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.329886 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.329978 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.329996 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.330020 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.330057 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:37Z","lastTransitionTime":"2026-01-29T16:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.433496 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.433578 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.433607 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.433710 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.433748 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:37Z","lastTransitionTime":"2026-01-29T16:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.537813 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.537896 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.537921 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.538002 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.538028 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:37Z","lastTransitionTime":"2026-01-29T16:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.640844 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.640884 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.640897 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.640913 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.640924 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:37Z","lastTransitionTime":"2026-01-29T16:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.743317 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.743359 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.743370 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.743385 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.743396 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:37Z","lastTransitionTime":"2026-01-29T16:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.847615 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.847697 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.847722 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.847754 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.847779 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:37Z","lastTransitionTime":"2026-01-29T16:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.879521 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs\") pod \"network-metrics-daemon-2w92b\" (UID: \"791456e8-8d95-4cdb-8fd1-d06a7586b328\") " pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:37 crc kubenswrapper[4714]: E0129 16:10:37.879763 4714 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 16:10:37 crc kubenswrapper[4714]: E0129 16:10:37.879908 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs podName:791456e8-8d95-4cdb-8fd1-d06a7586b328 nodeName:}" failed. No retries permitted until 2026-01-29 16:10:45.879871898 +0000 UTC m=+52.400373058 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs") pod "network-metrics-daemon-2w92b" (UID: "791456e8-8d95-4cdb-8fd1-d06a7586b328") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.951213 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.951279 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.951296 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.951321 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:37 crc kubenswrapper[4714]: I0129 16:10:37.951340 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:37Z","lastTransitionTime":"2026-01-29T16:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.055090 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.055154 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.055178 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.055210 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.055236 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:38Z","lastTransitionTime":"2026-01-29T16:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.153510 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 07:48:14.576587515 +0000 UTC Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.157774 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.157830 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.157850 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.157875 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.157892 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:38Z","lastTransitionTime":"2026-01-29T16:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.183484 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:38 crc kubenswrapper[4714]: E0129 16:10:38.183669 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.261141 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.261209 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.261233 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.261294 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.261317 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:38Z","lastTransitionTime":"2026-01-29T16:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.363492 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.363531 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.363541 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.363557 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.363568 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:38Z","lastTransitionTime":"2026-01-29T16:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.466918 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.467018 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.467036 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.467061 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.467078 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:38Z","lastTransitionTime":"2026-01-29T16:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.570374 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.570433 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.570449 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.570476 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.570494 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:38Z","lastTransitionTime":"2026-01-29T16:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.674158 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.674259 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.674277 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.674299 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.674316 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:38Z","lastTransitionTime":"2026-01-29T16:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.777424 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.777504 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.777525 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.777552 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.777570 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:38Z","lastTransitionTime":"2026-01-29T16:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.880322 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.880367 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.880383 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.880398 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.880409 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:38Z","lastTransitionTime":"2026-01-29T16:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.985432 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.985480 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.985492 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.985509 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:38 crc kubenswrapper[4714]: I0129 16:10:38.985519 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:38Z","lastTransitionTime":"2026-01-29T16:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.087693 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.087732 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.087743 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.087760 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.087772 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:39Z","lastTransitionTime":"2026-01-29T16:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.154201 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 14:21:24.805530221 +0000 UTC Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.183849 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.183875 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.183953 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:39 crc kubenswrapper[4714]: E0129 16:10:39.184084 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:39 crc kubenswrapper[4714]: E0129 16:10:39.184206 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:39 crc kubenswrapper[4714]: E0129 16:10:39.184330 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.191205 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.191235 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.191244 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.191257 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.191265 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:39Z","lastTransitionTime":"2026-01-29T16:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.294326 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.294385 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.294403 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.294425 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.294443 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:39Z","lastTransitionTime":"2026-01-29T16:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.397585 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.397727 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.397750 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.397777 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.397794 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:39Z","lastTransitionTime":"2026-01-29T16:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.500995 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.501056 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.501075 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.501099 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.501116 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:39Z","lastTransitionTime":"2026-01-29T16:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.604993 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.605059 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.605080 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.605111 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.605133 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:39Z","lastTransitionTime":"2026-01-29T16:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.708388 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.708709 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.708840 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.709006 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.709033 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:39Z","lastTransitionTime":"2026-01-29T16:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.811621 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.811682 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.811703 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.811728 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.811748 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:39Z","lastTransitionTime":"2026-01-29T16:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.915009 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.915097 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.915117 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.915140 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:39 crc kubenswrapper[4714]: I0129 16:10:39.915157 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:39Z","lastTransitionTime":"2026-01-29T16:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.017706 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.017792 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.017813 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.017842 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.017865 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:40Z","lastTransitionTime":"2026-01-29T16:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.121313 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.121414 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.121435 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.121514 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.121536 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:40Z","lastTransitionTime":"2026-01-29T16:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.154732 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 20:41:24.497319297 +0000 UTC Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.183767 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:40 crc kubenswrapper[4714]: E0129 16:10:40.184625 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.224765 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.224814 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.224830 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.224855 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.224874 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:40Z","lastTransitionTime":"2026-01-29T16:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.327794 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.327870 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.327892 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.327922 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.327972 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:40Z","lastTransitionTime":"2026-01-29T16:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.431545 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.431847 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.432036 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.432186 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.432401 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:40Z","lastTransitionTime":"2026-01-29T16:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.536597 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.536667 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.536693 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.536724 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.536748 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:40Z","lastTransitionTime":"2026-01-29T16:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.640299 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.640379 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.640413 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.640442 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.640463 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:40Z","lastTransitionTime":"2026-01-29T16:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.743369 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.743455 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.743481 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.743506 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.743524 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:40Z","lastTransitionTime":"2026-01-29T16:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.847355 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.847420 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.847441 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.847467 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.847484 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:40Z","lastTransitionTime":"2026-01-29T16:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.951378 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.951444 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.951462 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.951487 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:40 crc kubenswrapper[4714]: I0129 16:10:40.951506 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:40Z","lastTransitionTime":"2026-01-29T16:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.055588 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.055672 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.055699 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.055730 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.055755 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:41Z","lastTransitionTime":"2026-01-29T16:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.155045 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 13:32:15.739434184 +0000 UTC Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.158350 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.158410 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.158427 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.158453 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.158471 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:41Z","lastTransitionTime":"2026-01-29T16:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.184319 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.184386 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.184329 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:41 crc kubenswrapper[4714]: E0129 16:10:41.184513 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:41 crc kubenswrapper[4714]: E0129 16:10:41.184664 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:41 crc kubenswrapper[4714]: E0129 16:10:41.184795 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.263095 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.263198 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.263212 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.263233 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.263247 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:41Z","lastTransitionTime":"2026-01-29T16:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.367166 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.367234 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.367248 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.367272 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.367290 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:41Z","lastTransitionTime":"2026-01-29T16:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.458412 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.461631 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.461707 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.461725 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.461753 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.461777 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:41Z","lastTransitionTime":"2026-01-29T16:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.471719 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.478630 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: E0129 16:10:41.485655 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.492507 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.492576 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.492598 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.492628 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.492651 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:41Z","lastTransitionTime":"2026-01-29T16:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.504489 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: E0129 16:10:41.519004 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.524122 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.524216 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.524236 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.524261 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.524280 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:41Z","lastTransitionTime":"2026-01-29T16:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.531456 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: E0129 16:10:41.544856 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.548106 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.550630 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.550695 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.550714 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.550743 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.550764 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:41Z","lastTransitionTime":"2026-01-29T16:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.563438 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2w92b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791456e8-8d95-4cdb-8fd1-d06a7586b328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2w92b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: E0129 16:10:41.572685 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.577860 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.577916 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.577969 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.577999 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.578021 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:41Z","lastTransitionTime":"2026-01-29T16:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.582225 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.599408 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: E0129 16:10:41.600947 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: E0129 16:10:41.601324 4714 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.603340 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.603371 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.603386 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.603410 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.603425 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:41Z","lastTransitionTime":"2026-01-29T16:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.616419 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.636305 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.667735 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.689232 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.705717 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.705763 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.705780 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.705802 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.705818 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:41Z","lastTransitionTime":"2026-01-29T16:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.710751 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.728006 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.744342 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.759498 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.796335 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"message\\\":\\\"able:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 16:10:27.532704 6150 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-c9jhc\\\\nI0129 16:10:27.533048 6150 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0129 16:10:27.533063 6150 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-sbnkt_openshift-ovn-kubernetes(04b20f02-6c1e-4082-8233-8f06bda63195)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.809524 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.809578 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.809596 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.809621 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.809638 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:41Z","lastTransitionTime":"2026-01-29T16:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.818167 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2932c3bd-04c7-4494-8d43-03c4524a353f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78109e9e498f3d640934a2c45faf27133d32ece9e35e42ed48ddba720fa7a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc1d7b5f32fe8de4bfd5ba555838989293be538217424c86ea5cedadb8295f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tg8sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:41Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.913111 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.913559 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.913686 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.913795 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:41 crc kubenswrapper[4714]: I0129 16:10:41.913881 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:41Z","lastTransitionTime":"2026-01-29T16:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.018070 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.018139 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.018157 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.018183 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.018201 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:42Z","lastTransitionTime":"2026-01-29T16:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.121896 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.122021 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.122075 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.122133 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.122209 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:42Z","lastTransitionTime":"2026-01-29T16:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.155665 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 13:58:27.017021467 +0000 UTC Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.184355 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:42 crc kubenswrapper[4714]: E0129 16:10:42.184683 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.225741 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.225838 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.225866 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.225897 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.225920 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:42Z","lastTransitionTime":"2026-01-29T16:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.329486 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.329556 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.329577 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.329611 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.329633 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:42Z","lastTransitionTime":"2026-01-29T16:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.432135 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.432210 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.432240 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.432273 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.432295 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:42Z","lastTransitionTime":"2026-01-29T16:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.535411 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.535807 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.536053 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.536346 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.536743 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:42Z","lastTransitionTime":"2026-01-29T16:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.639770 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.639827 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.639845 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.639868 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.639885 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:42Z","lastTransitionTime":"2026-01-29T16:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.743416 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.743784 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.743988 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.744212 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.744423 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:42Z","lastTransitionTime":"2026-01-29T16:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.848209 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.848277 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.848295 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.848321 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.848340 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:42Z","lastTransitionTime":"2026-01-29T16:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.952044 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.952112 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.952129 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.952153 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:42 crc kubenswrapper[4714]: I0129 16:10:42.952174 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:42Z","lastTransitionTime":"2026-01-29T16:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.055037 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.055102 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.055127 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.055156 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.055180 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:43Z","lastTransitionTime":"2026-01-29T16:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.156377 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 11:13:54.491823218 +0000 UTC Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.158824 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.158877 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.158894 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.158918 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.158989 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:43Z","lastTransitionTime":"2026-01-29T16:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.183357 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.183423 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.183364 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:43 crc kubenswrapper[4714]: E0129 16:10:43.183551 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:43 crc kubenswrapper[4714]: E0129 16:10:43.183711 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:43 crc kubenswrapper[4714]: E0129 16:10:43.183830 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.262246 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.262321 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.262342 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.262367 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.262386 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:43Z","lastTransitionTime":"2026-01-29T16:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.366071 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.366144 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.366162 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.366189 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.366207 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:43Z","lastTransitionTime":"2026-01-29T16:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.469639 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.469699 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.469715 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.469747 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.469763 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:43Z","lastTransitionTime":"2026-01-29T16:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.572119 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.572186 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.572210 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.572240 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.572263 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:43Z","lastTransitionTime":"2026-01-29T16:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.675377 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.675424 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.675440 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.675463 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.675482 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:43Z","lastTransitionTime":"2026-01-29T16:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.781133 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.781190 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.781203 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.781224 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.781237 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:43Z","lastTransitionTime":"2026-01-29T16:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.884349 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.884411 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.884431 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.884456 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.884474 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:43Z","lastTransitionTime":"2026-01-29T16:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.987475 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.987530 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.987548 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.987570 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:43 crc kubenswrapper[4714]: I0129 16:10:43.987587 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:43Z","lastTransitionTime":"2026-01-29T16:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.091176 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.091243 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.091263 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.091287 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.091307 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:44Z","lastTransitionTime":"2026-01-29T16:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.157155 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 00:26:57.601562412 +0000 UTC Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.183809 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:44 crc kubenswrapper[4714]: E0129 16:10:44.184049 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.193616 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.193677 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.193699 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.193727 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.193749 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:44Z","lastTransitionTime":"2026-01-29T16:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.207701 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2932c3bd-04c7-4494-8d43-03c4524a353f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78109e9e498f3d640934a2c45faf27133d32ece9e35e42ed48ddba720fa7a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc1d7b5f32fe8de4bfd5ba555838989293be538217424c86ea5cedadb8295f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tg8sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:44Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.230902 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89403a5-379d-4c3f-a87f-8d2ed63ab368\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0bdc60ea5a5e188b2ed8894bb387084567b294c3a356fed01493d4bd6a7caf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6068173079b1d5cf7de92fe34bd0a4701b11a4dba7ae384f2d0e33f185656107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6565dc4c35c618a239bdc2be6dd45a9057e83573c92c3fc816350eb014c822c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:44Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.253189 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:44Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.269974 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:44Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.296755 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.296806 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.296825 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.296850 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.296866 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:44Z","lastTransitionTime":"2026-01-29T16:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.302813 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"message\\\":\\\"able:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 16:10:27.532704 6150 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-c9jhc\\\\nI0129 16:10:27.533048 6150 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0129 16:10:27.533063 6150 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-sbnkt_openshift-ovn-kubernetes(04b20f02-6c1e-4082-8233-8f06bda63195)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:44Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.320874 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2w92b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791456e8-8d95-4cdb-8fd1-d06a7586b328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2w92b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:44Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.340627 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:44Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.359771 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:44Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.382636 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:44Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.400243 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.400289 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.400305 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.400325 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.400341 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:44Z","lastTransitionTime":"2026-01-29T16:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.400504 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:44Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.423989 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:44Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.444191 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:44Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.460891 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:44Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.480593 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:44Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.497976 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:44Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.502873 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.502962 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.502983 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.503012 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.503033 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:44Z","lastTransitionTime":"2026-01-29T16:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.522628 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:44Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.546416 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:44Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.567719 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:44Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.606040 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.606104 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.606131 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.606162 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.606186 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:44Z","lastTransitionTime":"2026-01-29T16:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.709917 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.710032 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.710051 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.710081 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.710100 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:44Z","lastTransitionTime":"2026-01-29T16:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.812963 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.813023 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.813047 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.813079 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.813100 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:44Z","lastTransitionTime":"2026-01-29T16:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.915823 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.915907 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.915921 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.915975 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:44 crc kubenswrapper[4714]: I0129 16:10:44.915994 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:44Z","lastTransitionTime":"2026-01-29T16:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.019361 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.019424 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.019436 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.019456 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.019468 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:45Z","lastTransitionTime":"2026-01-29T16:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.122363 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.122426 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.122444 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.122468 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.122486 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:45Z","lastTransitionTime":"2026-01-29T16:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.157761 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 16:03:11.704903473 +0000 UTC Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.183417 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.183501 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.183438 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:45 crc kubenswrapper[4714]: E0129 16:10:45.183588 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:45 crc kubenswrapper[4714]: E0129 16:10:45.183693 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:45 crc kubenswrapper[4714]: E0129 16:10:45.183872 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.226659 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.226743 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.226763 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.227701 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.227796 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:45Z","lastTransitionTime":"2026-01-29T16:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.330304 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.330358 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.330371 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.330393 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.330409 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:45Z","lastTransitionTime":"2026-01-29T16:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.433517 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.433561 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.433574 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.433594 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.433604 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:45Z","lastTransitionTime":"2026-01-29T16:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.537793 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.537897 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.537908 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.537948 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.537961 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:45Z","lastTransitionTime":"2026-01-29T16:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.641120 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.641171 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.641188 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.641210 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.641227 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:45Z","lastTransitionTime":"2026-01-29T16:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.744057 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.744108 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.744122 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.744140 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.744151 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:45Z","lastTransitionTime":"2026-01-29T16:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.849406 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.849511 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.849531 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.849567 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.849589 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:45Z","lastTransitionTime":"2026-01-29T16:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.952671 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.952728 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.952751 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.952780 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.952805 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:45Z","lastTransitionTime":"2026-01-29T16:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:45 crc kubenswrapper[4714]: I0129 16:10:45.972011 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs\") pod \"network-metrics-daemon-2w92b\" (UID: \"791456e8-8d95-4cdb-8fd1-d06a7586b328\") " pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:45 crc kubenswrapper[4714]: E0129 16:10:45.972219 4714 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 16:10:45 crc kubenswrapper[4714]: E0129 16:10:45.972310 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs podName:791456e8-8d95-4cdb-8fd1-d06a7586b328 nodeName:}" failed. No retries permitted until 2026-01-29 16:11:01.97228666 +0000 UTC m=+68.492787780 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs") pod "network-metrics-daemon-2w92b" (UID: "791456e8-8d95-4cdb-8fd1-d06a7586b328") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.056578 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.056628 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.056641 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.056658 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.056670 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:46Z","lastTransitionTime":"2026-01-29T16:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.157985 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 05:08:13.085978649 +0000 UTC Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.160523 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.160583 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.160604 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.160628 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.160645 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:46Z","lastTransitionTime":"2026-01-29T16:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.183813 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:46 crc kubenswrapper[4714]: E0129 16:10:46.184400 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.184733 4714 scope.go:117] "RemoveContainer" containerID="5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.263778 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.263845 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.263863 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.263891 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.263914 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:46Z","lastTransitionTime":"2026-01-29T16:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.275504 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:10:46 crc kubenswrapper[4714]: E0129 16:10:46.275672 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:11:18.275640796 +0000 UTC m=+84.796141946 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.367114 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.367391 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.367523 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.367659 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.367775 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:46Z","lastTransitionTime":"2026-01-29T16:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.376799 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.376873 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.376921 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:46 crc kubenswrapper[4714]: E0129 16:10:46.377213 4714 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 16:10:46 crc kubenswrapper[4714]: E0129 16:10:46.377303 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 16:11:18.377279047 +0000 UTC m=+84.897780197 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 16:10:46 crc kubenswrapper[4714]: E0129 16:10:46.377345 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 16:10:46 crc kubenswrapper[4714]: E0129 16:10:46.377388 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 16:10:46 crc kubenswrapper[4714]: E0129 16:10:46.377484 4714 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:46 crc kubenswrapper[4714]: E0129 16:10:46.377497 4714 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 16:10:46 crc kubenswrapper[4714]: E0129 16:10:46.377584 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 16:11:18.377557765 +0000 UTC m=+84.898058915 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 16:10:46 crc kubenswrapper[4714]: E0129 16:10:46.377611 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 16:11:18.377599346 +0000 UTC m=+84.898100496 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:46 crc kubenswrapper[4714]: E0129 16:10:46.377642 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 16:10:46 crc kubenswrapper[4714]: E0129 16:10:46.377697 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 16:10:46 crc kubenswrapper[4714]: E0129 16:10:46.377728 4714 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:46 crc kubenswrapper[4714]: E0129 16:10:46.377824 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 16:11:18.377796982 +0000 UTC m=+84.898298142 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.377016 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.471761 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.471814 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.471833 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.471857 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.471874 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:46Z","lastTransitionTime":"2026-01-29T16:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.537282 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-sbnkt_04b20f02-6c1e-4082-8233-8f06bda63195/ovnkube-controller/1.log" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.541493 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerStarted","Data":"98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc"} Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.541907 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.564688 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:46Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.574977 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.575007 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.575019 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.575036 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.575048 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:46Z","lastTransitionTime":"2026-01-29T16:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.588275 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:46Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.616168 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:46Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.641981 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:46Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.674065 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2w92b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791456e8-8d95-4cdb-8fd1-d06a7586b328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2w92b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:46Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.678213 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.678259 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.678271 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.678290 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.678307 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:46Z","lastTransitionTime":"2026-01-29T16:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.701708 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:46Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.723419 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:46Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.740046 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:46Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.760986 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:46Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.783774 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.783822 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.783835 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.783855 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.783871 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:46Z","lastTransitionTime":"2026-01-29T16:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.791994 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:46Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.812115 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:46Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.840145 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:46Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.858517 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:46Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.873051 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89403a5-379d-4c3f-a87f-8d2ed63ab368\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0bdc60ea5a5e188b2ed8894bb387084567b294c3a356fed01493d4bd6a7caf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6068173079b1d5cf7de92fe34bd0a4701b11a4dba7ae384f2d0e33f185656107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6565dc4c35c618a239bdc2be6dd45a9057e83573c92c3fc816350eb014c822c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:46Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.887350 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.887390 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.887403 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.887435 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.887450 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:46Z","lastTransitionTime":"2026-01-29T16:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.892990 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:46Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.906197 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:46Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.928264 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"message\\\":\\\"able:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 16:10:27.532704 6150 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-c9jhc\\\\nI0129 16:10:27.533048 6150 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0129 16:10:27.533063 6150 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:46Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.940380 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2932c3bd-04c7-4494-8d43-03c4524a353f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78109e9e498f3d640934a2c45faf27133d32ece9e35e42ed48ddba720fa7a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc1d7b5f32fe8de4bfd5ba555838989293be538217424c86ea5cedadb8295f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tg8sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:46Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.990846 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.990905 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.990928 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.990982 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:46 crc kubenswrapper[4714]: I0129 16:10:46.991002 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:46Z","lastTransitionTime":"2026-01-29T16:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.094257 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.094309 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.094325 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.094344 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.094360 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:47Z","lastTransitionTime":"2026-01-29T16:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.158290 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 21:50:29.457942339 +0000 UTC Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.183735 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.183771 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.183801 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:47 crc kubenswrapper[4714]: E0129 16:10:47.183946 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:47 crc kubenswrapper[4714]: E0129 16:10:47.184022 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:47 crc kubenswrapper[4714]: E0129 16:10:47.184258 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.197865 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.197901 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.197912 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.197948 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.197961 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:47Z","lastTransitionTime":"2026-01-29T16:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.301827 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.301882 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.301893 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.301914 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.301944 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:47Z","lastTransitionTime":"2026-01-29T16:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.405639 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.405717 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.405732 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.405756 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.405771 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:47Z","lastTransitionTime":"2026-01-29T16:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.508551 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.508610 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.508627 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.508653 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.508672 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:47Z","lastTransitionTime":"2026-01-29T16:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.550121 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-sbnkt_04b20f02-6c1e-4082-8233-8f06bda63195/ovnkube-controller/2.log" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.551589 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-sbnkt_04b20f02-6c1e-4082-8233-8f06bda63195/ovnkube-controller/1.log" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.557550 4714 generic.go:334] "Generic (PLEG): container finished" podID="04b20f02-6c1e-4082-8233-8f06bda63195" containerID="98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc" exitCode=1 Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.557629 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerDied","Data":"98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc"} Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.557716 4714 scope.go:117] "RemoveContainer" containerID="5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.558965 4714 scope.go:117] "RemoveContainer" containerID="98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc" Jan 29 16:10:47 crc kubenswrapper[4714]: E0129 16:10:47.559332 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-sbnkt_openshift-ovn-kubernetes(04b20f02-6c1e-4082-8233-8f06bda63195)\"" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.583621 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:47Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.607132 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:47Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.611561 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.611613 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.611632 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.611655 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.611671 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:47Z","lastTransitionTime":"2026-01-29T16:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.625082 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:47Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.646352 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:47Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.679495 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:47Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.700543 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:47Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.715218 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.715288 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.715311 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.715342 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.715364 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:47Z","lastTransitionTime":"2026-01-29T16:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.722744 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:47Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.741460 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:47Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.760892 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89403a5-379d-4c3f-a87f-8d2ed63ab368\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0bdc60ea5a5e188b2ed8894bb387084567b294c3a356fed01493d4bd6a7caf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6068173079b1d5cf7de92fe34bd0a4701b11a4dba7ae384f2d0e33f185656107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6565dc4c35c618a239bdc2be6dd45a9057e83573c92c3fc816350eb014c822c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:47Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.782561 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:47Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.800682 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:47Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.818228 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.818572 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.818757 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.818983 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.819174 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:47Z","lastTransitionTime":"2026-01-29T16:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.832468 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5156bbcc6abbf40a22215f2642e81fed55f603d494e294703866350469208075\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"message\\\":\\\"able:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 16:10:27.532704 6150 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-c9jhc\\\\nI0129 16:10:27.533048 6150 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0129 16:10:27.533063 6150 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: fail\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:47Z\\\",\\\"message\\\":\\\"ing reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:10:47.314985 6375 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:10:47.315336 6375 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 16:10:47.315637 6375 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0129 16:10:47.315679 6375 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0129 16:10:47.315724 6375 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 16:10:47.315752 6375 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 16:10:47.315774 6375 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0129 16:10:47.315793 6375 factory.go:656] Stopping watch factory\\\\nI0129 16:10:47.315813 6375 ovnkube.go:599] Stopped ovnkube\\\\nI0129 16:10:47.315849 6375 handler.go:208] Removed *v1.Node event handler 2\\\\nI0129 16:10:47.315864 6375 handler.go:208] Removed *v1.Node event handler 7\\\\nI0129 16:10:47.315882 6375 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0129 16:10:47.315889 6375 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:47Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.850837 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2932c3bd-04c7-4494-8d43-03c4524a353f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78109e9e498f3d640934a2c45faf27133d32ece9e35e42ed48ddba720fa7a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc1d7b5f32fe8de4bfd5ba555838989293be538217424c86ea5cedadb8295f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tg8sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:47Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.868104 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:47Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.889001 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:47Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.913036 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:47Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.922525 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.922576 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.922588 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.922608 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.922620 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:47Z","lastTransitionTime":"2026-01-29T16:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.932674 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:47Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:47 crc kubenswrapper[4714]: I0129 16:10:47.951173 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2w92b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791456e8-8d95-4cdb-8fd1-d06a7586b328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2w92b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:47Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.025421 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.025493 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.025512 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.025562 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.025581 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:48Z","lastTransitionTime":"2026-01-29T16:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.128554 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.128669 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.128694 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.128725 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.128790 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:48Z","lastTransitionTime":"2026-01-29T16:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.159129 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 07:17:41.242893521 +0000 UTC Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.183602 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:48 crc kubenswrapper[4714]: E0129 16:10:48.183798 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.232354 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.232403 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.232420 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.232442 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.232461 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:48Z","lastTransitionTime":"2026-01-29T16:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.335894 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.336012 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.336037 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.336068 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.336088 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:48Z","lastTransitionTime":"2026-01-29T16:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.439142 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.439207 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.439227 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.439252 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.439269 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:48Z","lastTransitionTime":"2026-01-29T16:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.541806 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.541872 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.541897 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.541927 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.541983 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:48Z","lastTransitionTime":"2026-01-29T16:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.568193 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-sbnkt_04b20f02-6c1e-4082-8233-8f06bda63195/ovnkube-controller/2.log" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.572520 4714 scope.go:117] "RemoveContainer" containerID="98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc" Jan 29 16:10:48 crc kubenswrapper[4714]: E0129 16:10:48.573035 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-sbnkt_openshift-ovn-kubernetes(04b20f02-6c1e-4082-8233-8f06bda63195)\"" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.589846 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:48Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.609235 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:48Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.623059 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:48Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.643566 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:48Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.645478 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.645540 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.645557 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.645582 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.645600 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:48Z","lastTransitionTime":"2026-01-29T16:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.669007 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:48Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.687814 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:48Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.709471 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:48Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.724780 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:48Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.742849 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89403a5-379d-4c3f-a87f-8d2ed63ab368\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0bdc60ea5a5e188b2ed8894bb387084567b294c3a356fed01493d4bd6a7caf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6068173079b1d5cf7de92fe34bd0a4701b11a4dba7ae384f2d0e33f185656107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6565dc4c35c618a239bdc2be6dd45a9057e83573c92c3fc816350eb014c822c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:48Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.748217 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.748269 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.748283 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.748326 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.748338 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:48Z","lastTransitionTime":"2026-01-29T16:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.762172 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:48Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.777536 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:48Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.817850 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:47Z\\\",\\\"message\\\":\\\"ing reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:10:47.314985 6375 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:10:47.315336 6375 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 16:10:47.315637 6375 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0129 16:10:47.315679 6375 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0129 16:10:47.315724 6375 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 16:10:47.315752 6375 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 16:10:47.315774 6375 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0129 16:10:47.315793 6375 factory.go:656] Stopping watch factory\\\\nI0129 16:10:47.315813 6375 ovnkube.go:599] Stopped ovnkube\\\\nI0129 16:10:47.315849 6375 handler.go:208] Removed *v1.Node event handler 2\\\\nI0129 16:10:47.315864 6375 handler.go:208] Removed *v1.Node event handler 7\\\\nI0129 16:10:47.315882 6375 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0129 16:10:47.315889 6375 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-sbnkt_openshift-ovn-kubernetes(04b20f02-6c1e-4082-8233-8f06bda63195)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:48Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.836646 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2932c3bd-04c7-4494-8d43-03c4524a353f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78109e9e498f3d640934a2c45faf27133d32ece9e35e42ed48ddba720fa7a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc1d7b5f32fe8de4bfd5ba555838989293be538217424c86ea5cedadb8295f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tg8sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:48Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.850874 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.850918 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.850949 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.850969 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.850982 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:48Z","lastTransitionTime":"2026-01-29T16:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.855895 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:48Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.868784 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:48Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.880978 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:48Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.889443 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:48Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.900797 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2w92b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791456e8-8d95-4cdb-8fd1-d06a7586b328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2w92b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:48Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.953951 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.953983 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.953994 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.954008 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:48 crc kubenswrapper[4714]: I0129 16:10:48.954017 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:48Z","lastTransitionTime":"2026-01-29T16:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.056811 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.056872 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.056891 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.056912 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.056956 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:49Z","lastTransitionTime":"2026-01-29T16:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.159263 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 17:12:31.630409727 +0000 UTC Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.159714 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.159756 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.159766 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.159801 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.159812 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:49Z","lastTransitionTime":"2026-01-29T16:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.183698 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.183747 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.183896 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:49 crc kubenswrapper[4714]: E0129 16:10:49.184049 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:49 crc kubenswrapper[4714]: E0129 16:10:49.184235 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:49 crc kubenswrapper[4714]: E0129 16:10:49.184404 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.262711 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.262774 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.262791 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.262817 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.262835 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:49Z","lastTransitionTime":"2026-01-29T16:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.366053 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.366123 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.366141 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.366170 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.366187 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:49Z","lastTransitionTime":"2026-01-29T16:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.469682 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.469745 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.469771 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.469802 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.469828 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:49Z","lastTransitionTime":"2026-01-29T16:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.572404 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.572579 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.572653 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.572725 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.572754 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:49Z","lastTransitionTime":"2026-01-29T16:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.676227 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.676306 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.676324 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.676355 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.676375 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:49Z","lastTransitionTime":"2026-01-29T16:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.780025 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.780133 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.780152 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.780178 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.780194 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:49Z","lastTransitionTime":"2026-01-29T16:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.882891 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.882993 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.883012 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.883038 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.883057 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:49Z","lastTransitionTime":"2026-01-29T16:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.986476 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.986564 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.986582 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.986612 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:49 crc kubenswrapper[4714]: I0129 16:10:49.986634 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:49Z","lastTransitionTime":"2026-01-29T16:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.090041 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.090136 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.090155 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.090185 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.090205 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:50Z","lastTransitionTime":"2026-01-29T16:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.160146 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 07:14:44.941441217 +0000 UTC Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.183760 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:50 crc kubenswrapper[4714]: E0129 16:10:50.184262 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.195185 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.195265 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.195285 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.195313 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.195333 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:50Z","lastTransitionTime":"2026-01-29T16:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.298077 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.298132 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.298148 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.298172 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.298187 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:50Z","lastTransitionTime":"2026-01-29T16:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.401394 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.401467 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.401488 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.401519 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.401543 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:50Z","lastTransitionTime":"2026-01-29T16:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.505064 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.505118 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.505136 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.505161 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.505179 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:50Z","lastTransitionTime":"2026-01-29T16:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.608603 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.608676 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.608698 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.608729 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.608754 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:50Z","lastTransitionTime":"2026-01-29T16:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.711825 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.711877 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.711890 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.711910 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.711922 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:50Z","lastTransitionTime":"2026-01-29T16:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.815233 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.815368 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.815392 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.815415 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.815433 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:50Z","lastTransitionTime":"2026-01-29T16:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.919163 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.919257 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.919366 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.919392 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:50 crc kubenswrapper[4714]: I0129 16:10:50.919410 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:50Z","lastTransitionTime":"2026-01-29T16:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.021909 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.022026 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.022044 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.022068 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.022085 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:51Z","lastTransitionTime":"2026-01-29T16:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.125158 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.125269 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.125295 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.125325 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.125346 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:51Z","lastTransitionTime":"2026-01-29T16:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.161058 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 19:46:19.839056109 +0000 UTC Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.184001 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.184082 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.184021 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:51 crc kubenswrapper[4714]: E0129 16:10:51.184224 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:51 crc kubenswrapper[4714]: E0129 16:10:51.184393 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:51 crc kubenswrapper[4714]: E0129 16:10:51.184564 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.229139 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.229196 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.229218 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.229243 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.229261 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:51Z","lastTransitionTime":"2026-01-29T16:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.332180 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.332248 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.332265 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.332290 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.332309 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:51Z","lastTransitionTime":"2026-01-29T16:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.435428 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.435494 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.435514 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.435545 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.435567 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:51Z","lastTransitionTime":"2026-01-29T16:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.538164 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.538211 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.538228 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.538251 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.538268 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:51Z","lastTransitionTime":"2026-01-29T16:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.604645 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.604724 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.604749 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.604778 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.604801 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:51Z","lastTransitionTime":"2026-01-29T16:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:51 crc kubenswrapper[4714]: E0129 16:10:51.621567 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:51Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.626877 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.626925 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.627001 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.627023 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.627040 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:51Z","lastTransitionTime":"2026-01-29T16:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:51 crc kubenswrapper[4714]: E0129 16:10:51.647638 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:51Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.652208 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.652238 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.652249 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.652264 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.652276 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:51Z","lastTransitionTime":"2026-01-29T16:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:51 crc kubenswrapper[4714]: E0129 16:10:51.669640 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:51Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.674189 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.674238 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.674255 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.674277 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.674294 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:51Z","lastTransitionTime":"2026-01-29T16:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:51 crc kubenswrapper[4714]: E0129 16:10:51.694611 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:51Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.699675 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.699756 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.699774 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.699810 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.699837 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:51Z","lastTransitionTime":"2026-01-29T16:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:51 crc kubenswrapper[4714]: E0129 16:10:51.719606 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:51Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:51 crc kubenswrapper[4714]: E0129 16:10:51.719831 4714 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.722063 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.722122 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.722148 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.722176 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.722197 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:51Z","lastTransitionTime":"2026-01-29T16:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.825069 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.825145 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.825167 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.825195 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.825217 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:51Z","lastTransitionTime":"2026-01-29T16:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.928237 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.928329 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.928354 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.928387 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:51 crc kubenswrapper[4714]: I0129 16:10:51.928438 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:51Z","lastTransitionTime":"2026-01-29T16:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.031910 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.032003 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.032023 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.032049 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.032069 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:52Z","lastTransitionTime":"2026-01-29T16:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.135025 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.135454 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.135627 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.135798 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.136000 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:52Z","lastTransitionTime":"2026-01-29T16:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.161994 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 04:38:12.991351199 +0000 UTC Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.183598 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:52 crc kubenswrapper[4714]: E0129 16:10:52.184188 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.239681 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.239732 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.239749 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.239773 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.239792 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:52Z","lastTransitionTime":"2026-01-29T16:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.342862 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.342989 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.343016 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.343047 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.343072 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:52Z","lastTransitionTime":"2026-01-29T16:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.445443 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.445533 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.445617 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.445736 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.445784 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:52Z","lastTransitionTime":"2026-01-29T16:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.548911 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.548999 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.549014 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.549043 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.549059 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:52Z","lastTransitionTime":"2026-01-29T16:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.652311 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.652400 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.652436 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.652470 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.652493 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:52Z","lastTransitionTime":"2026-01-29T16:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.755551 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.755638 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.755658 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.755684 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.755702 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:52Z","lastTransitionTime":"2026-01-29T16:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.858765 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.858827 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.858844 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.858868 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.858887 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:52Z","lastTransitionTime":"2026-01-29T16:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.961144 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.961202 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.961213 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.961228 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:52 crc kubenswrapper[4714]: I0129 16:10:52.961239 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:52Z","lastTransitionTime":"2026-01-29T16:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.063849 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.063921 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.063974 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.064009 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.064075 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:53Z","lastTransitionTime":"2026-01-29T16:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.163502 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 22:47:23.895518493 +0000 UTC Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.166672 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.166735 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.166755 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.166812 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.166829 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:53Z","lastTransitionTime":"2026-01-29T16:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.183299 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.183392 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.183448 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:53 crc kubenswrapper[4714]: E0129 16:10:53.183608 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:53 crc kubenswrapper[4714]: E0129 16:10:53.183753 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:53 crc kubenswrapper[4714]: E0129 16:10:53.183991 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.269650 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.269717 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.269734 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.269762 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.269785 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:53Z","lastTransitionTime":"2026-01-29T16:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.373121 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.373189 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.373206 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.373231 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.373249 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:53Z","lastTransitionTime":"2026-01-29T16:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.476400 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.476481 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.476505 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.476540 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.476562 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:53Z","lastTransitionTime":"2026-01-29T16:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.579300 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.579357 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.579372 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.579406 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.579418 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:53Z","lastTransitionTime":"2026-01-29T16:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.683311 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.683371 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.683383 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.683406 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.683421 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:53Z","lastTransitionTime":"2026-01-29T16:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.786211 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.786257 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.786269 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.786285 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.786299 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:53Z","lastTransitionTime":"2026-01-29T16:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.889034 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.889113 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.889141 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.889175 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.889200 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:53Z","lastTransitionTime":"2026-01-29T16:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.991071 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.991123 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.991132 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.991145 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:53 crc kubenswrapper[4714]: I0129 16:10:53.991154 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:53Z","lastTransitionTime":"2026-01-29T16:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.094838 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.094903 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.094912 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.094949 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.094962 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:54Z","lastTransitionTime":"2026-01-29T16:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.164503 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 15:16:39.879751955 +0000 UTC Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.183624 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:54 crc kubenswrapper[4714]: E0129 16:10:54.183771 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.197789 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.198127 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.198265 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.198398 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.198498 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:54Z","lastTransitionTime":"2026-01-29T16:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.199703 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2w92b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791456e8-8d95-4cdb-8fd1-d06a7586b328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2w92b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:54Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.220005 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:54Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.237211 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:54Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.255138 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:54Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.272657 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:54Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.296679 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:54Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.301170 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.301231 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.301249 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.301279 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.301299 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:54Z","lastTransitionTime":"2026-01-29T16:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.311072 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:54Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.323881 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:54Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.339049 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:54Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.353521 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:54Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.378773 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:54Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.395211 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:54Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.403451 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.403519 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.403534 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.403571 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.403607 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:54Z","lastTransitionTime":"2026-01-29T16:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.417080 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:54Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.436662 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2932c3bd-04c7-4494-8d43-03c4524a353f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78109e9e498f3d640934a2c45faf27133d32ece9e35e42ed48ddba720fa7a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc1d7b5f32fe8de4bfd5ba555838989293be538217424c86ea5cedadb8295f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tg8sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:54Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.456248 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89403a5-379d-4c3f-a87f-8d2ed63ab368\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0bdc60ea5a5e188b2ed8894bb387084567b294c3a356fed01493d4bd6a7caf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6068173079b1d5cf7de92fe34bd0a4701b11a4dba7ae384f2d0e33f185656107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6565dc4c35c618a239bdc2be6dd45a9057e83573c92c3fc816350eb014c822c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:54Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.475082 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:54Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.490503 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:54Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.505973 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.506054 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.506080 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.506124 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.506149 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:54Z","lastTransitionTime":"2026-01-29T16:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.518631 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:47Z\\\",\\\"message\\\":\\\"ing reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:10:47.314985 6375 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:10:47.315336 6375 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 16:10:47.315637 6375 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0129 16:10:47.315679 6375 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0129 16:10:47.315724 6375 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 16:10:47.315752 6375 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 16:10:47.315774 6375 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0129 16:10:47.315793 6375 factory.go:656] Stopping watch factory\\\\nI0129 16:10:47.315813 6375 ovnkube.go:599] Stopped ovnkube\\\\nI0129 16:10:47.315849 6375 handler.go:208] Removed *v1.Node event handler 2\\\\nI0129 16:10:47.315864 6375 handler.go:208] Removed *v1.Node event handler 7\\\\nI0129 16:10:47.315882 6375 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0129 16:10:47.315889 6375 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-sbnkt_openshift-ovn-kubernetes(04b20f02-6c1e-4082-8233-8f06bda63195)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:10:54Z is after 2025-08-24T17:21:41Z" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.609395 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.609454 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.609472 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.609494 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.609892 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:54Z","lastTransitionTime":"2026-01-29T16:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.712656 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.712886 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.712896 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.712911 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.712922 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:54Z","lastTransitionTime":"2026-01-29T16:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.816450 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.816820 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.817004 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.817163 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.817305 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:54Z","lastTransitionTime":"2026-01-29T16:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.920627 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.920695 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.920718 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.920745 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:54 crc kubenswrapper[4714]: I0129 16:10:54.920766 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:54Z","lastTransitionTime":"2026-01-29T16:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.024299 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.024370 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.024404 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.024425 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.024442 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:55Z","lastTransitionTime":"2026-01-29T16:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.126240 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.126278 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.126290 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.126305 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.126318 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:55Z","lastTransitionTime":"2026-01-29T16:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.164850 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 02:35:30.42743811 +0000 UTC Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.183213 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:55 crc kubenswrapper[4714]: E0129 16:10:55.183321 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.183374 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:55 crc kubenswrapper[4714]: E0129 16:10:55.183551 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.183570 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:55 crc kubenswrapper[4714]: E0129 16:10:55.183717 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.229061 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.229106 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.229122 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.229147 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.229164 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:55Z","lastTransitionTime":"2026-01-29T16:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.332310 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.332710 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.332866 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.333063 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.333238 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:55Z","lastTransitionTime":"2026-01-29T16:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.435731 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.435779 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.435790 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.435808 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.435822 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:55Z","lastTransitionTime":"2026-01-29T16:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.539473 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.539543 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.539561 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.539586 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.539604 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:55Z","lastTransitionTime":"2026-01-29T16:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.642810 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.643129 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.643209 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.643291 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.643399 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:55Z","lastTransitionTime":"2026-01-29T16:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.746227 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.746289 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.746306 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.746327 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.746345 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:55Z","lastTransitionTime":"2026-01-29T16:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.849433 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.849502 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.849527 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.849556 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.849580 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:55Z","lastTransitionTime":"2026-01-29T16:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.953329 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.953370 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.953382 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.953402 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:55 crc kubenswrapper[4714]: I0129 16:10:55.953414 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:55Z","lastTransitionTime":"2026-01-29T16:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.057638 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.057681 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.057695 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.057716 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.057729 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:56Z","lastTransitionTime":"2026-01-29T16:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.160686 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.160739 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.160757 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.160786 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.160804 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:56Z","lastTransitionTime":"2026-01-29T16:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.263266 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.263310 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.263326 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.263348 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.263364 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:56Z","lastTransitionTime":"2026-01-29T16:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.366974 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.367034 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.367053 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.367074 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.367091 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:56Z","lastTransitionTime":"2026-01-29T16:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.469825 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.469872 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.469889 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.469913 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.469959 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:56Z","lastTransitionTime":"2026-01-29T16:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.572362 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.572415 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.572432 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.572451 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.572467 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:56Z","lastTransitionTime":"2026-01-29T16:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.675038 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.675071 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.675083 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.675100 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.675112 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:56Z","lastTransitionTime":"2026-01-29T16:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.779016 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.779053 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.779062 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.779079 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.779089 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:56Z","lastTransitionTime":"2026-01-29T16:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.845475 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 08:33:53.939539328 +0000 UTC Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.846355 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.846465 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.846516 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:56 crc kubenswrapper[4714]: E0129 16:10:56.846508 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.846465 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:56 crc kubenswrapper[4714]: E0129 16:10:56.846664 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:10:56 crc kubenswrapper[4714]: E0129 16:10:56.846792 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:56 crc kubenswrapper[4714]: E0129 16:10:56.846903 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.882358 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.882425 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.882441 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.882468 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.882486 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:56Z","lastTransitionTime":"2026-01-29T16:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.985733 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.985815 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.985834 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.985864 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:56 crc kubenswrapper[4714]: I0129 16:10:56.985885 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:56Z","lastTransitionTime":"2026-01-29T16:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.088991 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.089048 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.089067 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.089094 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.089114 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:57Z","lastTransitionTime":"2026-01-29T16:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.192730 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.192802 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.192823 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.192848 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.192866 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:57Z","lastTransitionTime":"2026-01-29T16:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.296034 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.296098 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.296118 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.296147 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.296165 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:57Z","lastTransitionTime":"2026-01-29T16:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.399433 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.399508 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.399528 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.399553 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.399571 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:57Z","lastTransitionTime":"2026-01-29T16:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.502722 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.502778 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.502794 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.502816 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.502833 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:57Z","lastTransitionTime":"2026-01-29T16:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.604672 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.604718 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.604730 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.604749 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.604763 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:57Z","lastTransitionTime":"2026-01-29T16:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.707482 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.707553 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.707573 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.707598 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.707616 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:57Z","lastTransitionTime":"2026-01-29T16:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.811345 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.811415 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.811433 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.811459 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.811481 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:57Z","lastTransitionTime":"2026-01-29T16:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.845664 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 06:35:04.24250712 +0000 UTC Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.914798 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.914843 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.914854 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.914871 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:57 crc kubenswrapper[4714]: I0129 16:10:57.914883 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:57Z","lastTransitionTime":"2026-01-29T16:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.018175 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.018238 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.018256 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.018280 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.018298 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:58Z","lastTransitionTime":"2026-01-29T16:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.121524 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.121605 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.121624 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.121649 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.121665 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:58Z","lastTransitionTime":"2026-01-29T16:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.184108 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.184186 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.184249 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:10:58 crc kubenswrapper[4714]: E0129 16:10:58.184386 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:10:58 crc kubenswrapper[4714]: E0129 16:10:58.184567 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:10:58 crc kubenswrapper[4714]: E0129 16:10:58.184665 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.224511 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.224552 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.224565 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.224605 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.224618 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:58Z","lastTransitionTime":"2026-01-29T16:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.328255 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.328295 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.328331 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.328348 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.328361 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:58Z","lastTransitionTime":"2026-01-29T16:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.431864 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.432157 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.432215 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.432240 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.432601 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:58Z","lastTransitionTime":"2026-01-29T16:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.535553 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.535599 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.535615 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.535638 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.535655 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:58Z","lastTransitionTime":"2026-01-29T16:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.638494 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.638559 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.638577 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.638601 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.638619 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:58Z","lastTransitionTime":"2026-01-29T16:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.741649 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.741707 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.741724 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.741749 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.741767 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:58Z","lastTransitionTime":"2026-01-29T16:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.845680 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.845734 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.845753 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.845777 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.845796 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:58Z","lastTransitionTime":"2026-01-29T16:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.845917 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 14:08:11.895414663 +0000 UTC Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.949298 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.949379 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.949416 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.949438 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:58 crc kubenswrapper[4714]: I0129 16:10:58.949449 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:58Z","lastTransitionTime":"2026-01-29T16:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.053463 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.053524 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.053541 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.053572 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.053590 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:59Z","lastTransitionTime":"2026-01-29T16:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.156979 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.157030 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.157048 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.157072 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.157103 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:59Z","lastTransitionTime":"2026-01-29T16:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.183875 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:10:59 crc kubenswrapper[4714]: E0129 16:10:59.184039 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.260382 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.260440 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.260460 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.260487 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.260506 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:59Z","lastTransitionTime":"2026-01-29T16:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.363367 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.363411 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.363423 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.363441 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.363454 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:59Z","lastTransitionTime":"2026-01-29T16:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.466706 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.466777 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.466796 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.466828 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.466846 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:59Z","lastTransitionTime":"2026-01-29T16:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.569848 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.569911 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.570192 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.570235 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.570256 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:59Z","lastTransitionTime":"2026-01-29T16:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.672788 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.672830 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.672843 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.672862 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.672874 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:59Z","lastTransitionTime":"2026-01-29T16:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.776531 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.776596 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.776615 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.776640 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.776658 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:59Z","lastTransitionTime":"2026-01-29T16:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.846229 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 16:34:18.566218485 +0000 UTC Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.888530 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.888575 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.888590 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.888612 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.888627 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:59Z","lastTransitionTime":"2026-01-29T16:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.992127 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.992201 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.992221 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.992247 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:10:59 crc kubenswrapper[4714]: I0129 16:10:59.992269 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:10:59Z","lastTransitionTime":"2026-01-29T16:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.095173 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.095235 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.095252 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.095280 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.095297 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:00Z","lastTransitionTime":"2026-01-29T16:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.183369 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.183414 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:00 crc kubenswrapper[4714]: E0129 16:11:00.183544 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.183564 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:00 crc kubenswrapper[4714]: E0129 16:11:00.183660 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:00 crc kubenswrapper[4714]: E0129 16:11:00.183736 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.201112 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.201172 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.201195 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.201220 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.201237 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:00Z","lastTransitionTime":"2026-01-29T16:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.303634 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.303667 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.303675 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.303688 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.303697 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:00Z","lastTransitionTime":"2026-01-29T16:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.406414 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.406468 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.406486 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.406510 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.406525 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:00Z","lastTransitionTime":"2026-01-29T16:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.522072 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.522134 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.522155 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.522180 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.522198 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:00Z","lastTransitionTime":"2026-01-29T16:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.624284 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.624323 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.624333 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.624348 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.624357 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:00Z","lastTransitionTime":"2026-01-29T16:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.726878 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.726969 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.726992 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.727023 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.727048 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:00Z","lastTransitionTime":"2026-01-29T16:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.830247 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.830301 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.830316 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.830340 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.830353 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:00Z","lastTransitionTime":"2026-01-29T16:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.846719 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 07:15:56.643904055 +0000 UTC Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.932903 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.932953 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.932965 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.932981 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:00 crc kubenswrapper[4714]: I0129 16:11:00.932992 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:00Z","lastTransitionTime":"2026-01-29T16:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.035895 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.035980 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.036000 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.036025 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.036042 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:01Z","lastTransitionTime":"2026-01-29T16:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.139651 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.139724 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.139751 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.139784 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.139806 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:01Z","lastTransitionTime":"2026-01-29T16:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.183320 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:01 crc kubenswrapper[4714]: E0129 16:11:01.183699 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.183892 4714 scope.go:117] "RemoveContainer" containerID="98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc" Jan 29 16:11:01 crc kubenswrapper[4714]: E0129 16:11:01.184052 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-sbnkt_openshift-ovn-kubernetes(04b20f02-6c1e-4082-8233-8f06bda63195)\"" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.242110 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.242142 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.242150 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.242164 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.242173 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:01Z","lastTransitionTime":"2026-01-29T16:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.345556 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.345600 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.345608 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.345625 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.345634 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:01Z","lastTransitionTime":"2026-01-29T16:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.447781 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.447831 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.447843 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.447864 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.447877 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:01Z","lastTransitionTime":"2026-01-29T16:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.550495 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.550553 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.550565 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.550584 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.550602 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:01Z","lastTransitionTime":"2026-01-29T16:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.654021 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.654064 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.654077 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.654094 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.654105 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:01Z","lastTransitionTime":"2026-01-29T16:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.756209 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.756247 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.756272 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.756287 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.756297 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:01Z","lastTransitionTime":"2026-01-29T16:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.847573 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 19:17:37.83068583 +0000 UTC Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.858339 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.858410 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.858418 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.858435 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.858444 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:01Z","lastTransitionTime":"2026-01-29T16:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.961636 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.961693 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.961709 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.961731 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.961744 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:01Z","lastTransitionTime":"2026-01-29T16:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.963024 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.963064 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.963081 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.963098 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.963112 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:01Z","lastTransitionTime":"2026-01-29T16:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:01 crc kubenswrapper[4714]: E0129 16:11:01.977630 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:01Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.980592 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.980628 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.980637 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.980654 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:01 crc kubenswrapper[4714]: I0129 16:11:01.980667 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:01Z","lastTransitionTime":"2026-01-29T16:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:01 crc kubenswrapper[4714]: E0129 16:11:01.997581 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:01Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.001971 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.002032 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.002049 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.002105 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.002121 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:02Z","lastTransitionTime":"2026-01-29T16:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.008646 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs\") pod \"network-metrics-daemon-2w92b\" (UID: \"791456e8-8d95-4cdb-8fd1-d06a7586b328\") " pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:02 crc kubenswrapper[4714]: E0129 16:11:02.008794 4714 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 16:11:02 crc kubenswrapper[4714]: E0129 16:11:02.008856 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs podName:791456e8-8d95-4cdb-8fd1-d06a7586b328 nodeName:}" failed. No retries permitted until 2026-01-29 16:11:34.008840376 +0000 UTC m=+100.529341506 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs") pod "network-metrics-daemon-2w92b" (UID: "791456e8-8d95-4cdb-8fd1-d06a7586b328") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 16:11:02 crc kubenswrapper[4714]: E0129 16:11:02.018794 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:02Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.022203 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.022252 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.022265 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.022284 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.022295 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:02Z","lastTransitionTime":"2026-01-29T16:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:02 crc kubenswrapper[4714]: E0129 16:11:02.036723 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:02Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.039804 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.039841 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.039853 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.039875 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.039887 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:02Z","lastTransitionTime":"2026-01-29T16:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:02 crc kubenswrapper[4714]: E0129 16:11:02.054736 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:02Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:02 crc kubenswrapper[4714]: E0129 16:11:02.054916 4714 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.064041 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.064085 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.064095 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.064115 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.064132 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:02Z","lastTransitionTime":"2026-01-29T16:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.168405 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.168463 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.168482 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.168503 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.168522 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:02Z","lastTransitionTime":"2026-01-29T16:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.183989 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.184031 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.183999 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:02 crc kubenswrapper[4714]: E0129 16:11:02.184142 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:02 crc kubenswrapper[4714]: E0129 16:11:02.184356 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:02 crc kubenswrapper[4714]: E0129 16:11:02.184421 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.271244 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.271330 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.271348 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.271367 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.271383 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:02Z","lastTransitionTime":"2026-01-29T16:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.373297 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.373335 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.373352 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.373371 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.373387 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:02Z","lastTransitionTime":"2026-01-29T16:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.475905 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.475988 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.476002 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.476021 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.476036 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:02Z","lastTransitionTime":"2026-01-29T16:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.579949 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.580002 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.580015 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.580035 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.580046 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:02Z","lastTransitionTime":"2026-01-29T16:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.683697 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.683745 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.683756 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.683781 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.683795 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:02Z","lastTransitionTime":"2026-01-29T16:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.787644 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.787696 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.787715 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.787740 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.787760 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:02Z","lastTransitionTime":"2026-01-29T16:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.847982 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 04:32:47.95244453 +0000 UTC Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.890295 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.890348 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.890365 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.890388 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.890407 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:02Z","lastTransitionTime":"2026-01-29T16:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.993022 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.993059 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.993069 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.993083 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:02 crc kubenswrapper[4714]: I0129 16:11:02.993092 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:02Z","lastTransitionTime":"2026-01-29T16:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.095545 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.095592 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.095606 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.095623 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.095676 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:03Z","lastTransitionTime":"2026-01-29T16:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.183660 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:03 crc kubenswrapper[4714]: E0129 16:11:03.183793 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.198820 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.198853 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.198863 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.198876 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.198885 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:03Z","lastTransitionTime":"2026-01-29T16:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.301869 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.301909 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.301922 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.301968 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.301983 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:03Z","lastTransitionTime":"2026-01-29T16:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.405161 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.405198 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.405208 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.405223 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.405235 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:03Z","lastTransitionTime":"2026-01-29T16:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.508391 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.508448 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.508461 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.508484 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.508503 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:03Z","lastTransitionTime":"2026-01-29T16:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.611260 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.611303 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.611315 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.611333 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.611345 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:03Z","lastTransitionTime":"2026-01-29T16:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.714714 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.714811 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.714832 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.714869 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.714890 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:03Z","lastTransitionTime":"2026-01-29T16:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.818284 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.818383 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.818569 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.818594 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.818608 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:03Z","lastTransitionTime":"2026-01-29T16:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.848870 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 01:27:13.90868659 +0000 UTC Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.922531 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.922578 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.922592 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.922608 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:03 crc kubenswrapper[4714]: I0129 16:11:03.922620 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:03Z","lastTransitionTime":"2026-01-29T16:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.025607 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.025651 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.025663 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.025682 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.025695 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:04Z","lastTransitionTime":"2026-01-29T16:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.129404 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.129462 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.129478 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.129505 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.129522 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:04Z","lastTransitionTime":"2026-01-29T16:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.183587 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:04 crc kubenswrapper[4714]: E0129 16:11:04.183733 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.183614 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:04 crc kubenswrapper[4714]: E0129 16:11:04.183808 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.183595 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:04 crc kubenswrapper[4714]: E0129 16:11:04.184097 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.201438 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.218332 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.232136 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.232214 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.232248 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.232273 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.232287 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:04Z","lastTransitionTime":"2026-01-29T16:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.232685 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.248360 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2w92b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791456e8-8d95-4cdb-8fd1-d06a7586b328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2w92b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.269344 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.284746 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.301125 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.313741 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.330101 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.335139 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.335193 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.335204 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.335227 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.335239 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:04Z","lastTransitionTime":"2026-01-29T16:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.364142 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.385351 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.406332 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.423619 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.438339 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.438397 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.438414 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.438438 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.438454 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:04Z","lastTransitionTime":"2026-01-29T16:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.442091 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.459347 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.479344 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:47Z\\\",\\\"message\\\":\\\"ing reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:10:47.314985 6375 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:10:47.315336 6375 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 16:10:47.315637 6375 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0129 16:10:47.315679 6375 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0129 16:10:47.315724 6375 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 16:10:47.315752 6375 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 16:10:47.315774 6375 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0129 16:10:47.315793 6375 factory.go:656] Stopping watch factory\\\\nI0129 16:10:47.315813 6375 ovnkube.go:599] Stopped ovnkube\\\\nI0129 16:10:47.315849 6375 handler.go:208] Removed *v1.Node event handler 2\\\\nI0129 16:10:47.315864 6375 handler.go:208] Removed *v1.Node event handler 7\\\\nI0129 16:10:47.315882 6375 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0129 16:10:47.315889 6375 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-sbnkt_openshift-ovn-kubernetes(04b20f02-6c1e-4082-8233-8f06bda63195)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.492557 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2932c3bd-04c7-4494-8d43-03c4524a353f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78109e9e498f3d640934a2c45faf27133d32ece9e35e42ed48ddba720fa7a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc1d7b5f32fe8de4bfd5ba555838989293be538217424c86ea5cedadb8295f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tg8sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.506316 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89403a5-379d-4c3f-a87f-8d2ed63ab368\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0bdc60ea5a5e188b2ed8894bb387084567b294c3a356fed01493d4bd6a7caf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6068173079b1d5cf7de92fe34bd0a4701b11a4dba7ae384f2d0e33f185656107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6565dc4c35c618a239bdc2be6dd45a9057e83573c92c3fc816350eb014c822c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.540919 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.540986 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.540996 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.541017 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.541031 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:04Z","lastTransitionTime":"2026-01-29T16:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.629695 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b2ttm_89560008-8bdc-4640-af11-681d825e69d4/kube-multus/0.log" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.629777 4714 generic.go:334] "Generic (PLEG): container finished" podID="89560008-8bdc-4640-af11-681d825e69d4" containerID="a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a" exitCode=1 Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.629836 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b2ttm" event={"ID":"89560008-8bdc-4640-af11-681d825e69d4","Type":"ContainerDied","Data":"a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a"} Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.630523 4714 scope.go:117] "RemoveContainer" containerID="a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.642747 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.642790 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.642802 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.642818 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.642828 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:04Z","lastTransitionTime":"2026-01-29T16:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.647137 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.664666 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.679569 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.691116 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:11:03Z\\\",\\\"message\\\":\\\"2026-01-29T16:10:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_27c11876-c5ae-4d35-b720-78ab09bfac0c\\\\n2026-01-29T16:10:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_27c11876-c5ae-4d35-b720-78ab09bfac0c to /host/opt/cni/bin/\\\\n2026-01-29T16:10:18Z [verbose] multus-daemon started\\\\n2026-01-29T16:10:18Z [verbose] Readiness Indicator file check\\\\n2026-01-29T16:11:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.717023 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.729919 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.743428 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.745506 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.745558 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.745572 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.745590 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.745602 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:04Z","lastTransitionTime":"2026-01-29T16:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.753701 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.764952 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89403a5-379d-4c3f-a87f-8d2ed63ab368\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0bdc60ea5a5e188b2ed8894bb387084567b294c3a356fed01493d4bd6a7caf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6068173079b1d5cf7de92fe34bd0a4701b11a4dba7ae384f2d0e33f185656107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6565dc4c35c618a239bdc2be6dd45a9057e83573c92c3fc816350eb014c822c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.776318 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.786862 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.809149 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:47Z\\\",\\\"message\\\":\\\"ing reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:10:47.314985 6375 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:10:47.315336 6375 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 16:10:47.315637 6375 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0129 16:10:47.315679 6375 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0129 16:10:47.315724 6375 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 16:10:47.315752 6375 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 16:10:47.315774 6375 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0129 16:10:47.315793 6375 factory.go:656] Stopping watch factory\\\\nI0129 16:10:47.315813 6375 ovnkube.go:599] Stopped ovnkube\\\\nI0129 16:10:47.315849 6375 handler.go:208] Removed *v1.Node event handler 2\\\\nI0129 16:10:47.315864 6375 handler.go:208] Removed *v1.Node event handler 7\\\\nI0129 16:10:47.315882 6375 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0129 16:10:47.315889 6375 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-sbnkt_openshift-ovn-kubernetes(04b20f02-6c1e-4082-8233-8f06bda63195)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.820639 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2932c3bd-04c7-4494-8d43-03c4524a353f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78109e9e498f3d640934a2c45faf27133d32ece9e35e42ed48ddba720fa7a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc1d7b5f32fe8de4bfd5ba555838989293be538217424c86ea5cedadb8295f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tg8sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.833138 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.847040 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.848447 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.848497 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.848513 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.848531 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.848544 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:04Z","lastTransitionTime":"2026-01-29T16:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.849046 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 08:13:09.765681762 +0000 UTC Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.863585 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.876481 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.889693 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2w92b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791456e8-8d95-4cdb-8fd1-d06a7586b328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2w92b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:04Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.951264 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.951309 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.951319 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.951343 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:04 crc kubenswrapper[4714]: I0129 16:11:04.951354 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:04Z","lastTransitionTime":"2026-01-29T16:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.054820 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.054873 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.054886 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.054907 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.054923 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:05Z","lastTransitionTime":"2026-01-29T16:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.157989 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.158035 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.158050 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.158071 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.158085 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:05Z","lastTransitionTime":"2026-01-29T16:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.183466 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:05 crc kubenswrapper[4714]: E0129 16:11:05.183630 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.261791 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.261857 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.261876 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.261900 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.261917 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:05Z","lastTransitionTime":"2026-01-29T16:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.365254 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.365329 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.365341 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.365365 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.365378 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:05Z","lastTransitionTime":"2026-01-29T16:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.468189 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.468272 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.468292 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.468318 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.468331 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:05Z","lastTransitionTime":"2026-01-29T16:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.571900 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.571953 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.571965 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.571979 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.571987 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:05Z","lastTransitionTime":"2026-01-29T16:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.636504 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b2ttm_89560008-8bdc-4640-af11-681d825e69d4/kube-multus/0.log" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.636571 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b2ttm" event={"ID":"89560008-8bdc-4640-af11-681d825e69d4","Type":"ContainerStarted","Data":"c747ed61a18e27d63630395860ce896242426b1ce46ea5f9d00534b808804a58"} Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.658923 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:05Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.675515 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.675566 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.675582 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.675603 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.675619 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:05Z","lastTransitionTime":"2026-01-29T16:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.685482 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:05Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.701056 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:05Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.719835 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:05Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.735190 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2932c3bd-04c7-4494-8d43-03c4524a353f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78109e9e498f3d640934a2c45faf27133d32ece9e35e42ed48ddba720fa7a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc1d7b5f32fe8de4bfd5ba555838989293be538217424c86ea5cedadb8295f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tg8sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:05Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.751766 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89403a5-379d-4c3f-a87f-8d2ed63ab368\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0bdc60ea5a5e188b2ed8894bb387084567b294c3a356fed01493d4bd6a7caf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6068173079b1d5cf7de92fe34bd0a4701b11a4dba7ae384f2d0e33f185656107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6565dc4c35c618a239bdc2be6dd45a9057e83573c92c3fc816350eb014c822c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:05Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.766982 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:05Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.778713 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.778764 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.778779 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.778798 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.778812 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:05Z","lastTransitionTime":"2026-01-29T16:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.782683 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:05Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.805899 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:47Z\\\",\\\"message\\\":\\\"ing reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:10:47.314985 6375 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:10:47.315336 6375 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 16:10:47.315637 6375 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0129 16:10:47.315679 6375 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0129 16:10:47.315724 6375 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 16:10:47.315752 6375 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 16:10:47.315774 6375 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0129 16:10:47.315793 6375 factory.go:656] Stopping watch factory\\\\nI0129 16:10:47.315813 6375 ovnkube.go:599] Stopped ovnkube\\\\nI0129 16:10:47.315849 6375 handler.go:208] Removed *v1.Node event handler 2\\\\nI0129 16:10:47.315864 6375 handler.go:208] Removed *v1.Node event handler 7\\\\nI0129 16:10:47.315882 6375 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0129 16:10:47.315889 6375 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-sbnkt_openshift-ovn-kubernetes(04b20f02-6c1e-4082-8233-8f06bda63195)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:05Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.821229 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2w92b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791456e8-8d95-4cdb-8fd1-d06a7586b328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2w92b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:05Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.834913 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:05Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.847030 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:05Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.849285 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 06:37:59.96683529 +0000 UTC Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.863585 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:05Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.877113 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:05Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.881427 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.881478 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.881489 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.881504 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.881513 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:05Z","lastTransitionTime":"2026-01-29T16:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.890687 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:05Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.906648 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:05Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.924037 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:05Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.941414 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c747ed61a18e27d63630395860ce896242426b1ce46ea5f9d00534b808804a58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:11:03Z\\\",\\\"message\\\":\\\"2026-01-29T16:10:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_27c11876-c5ae-4d35-b720-78ab09bfac0c\\\\n2026-01-29T16:10:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_27c11876-c5ae-4d35-b720-78ab09bfac0c to /host/opt/cni/bin/\\\\n2026-01-29T16:10:18Z [verbose] multus-daemon started\\\\n2026-01-29T16:10:18Z [verbose] Readiness Indicator file check\\\\n2026-01-29T16:11:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:05Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.985077 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.985142 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.985159 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.985180 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:05 crc kubenswrapper[4714]: I0129 16:11:05.985198 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:05Z","lastTransitionTime":"2026-01-29T16:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.087830 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.087876 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.087888 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.087903 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.087915 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:06Z","lastTransitionTime":"2026-01-29T16:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.184063 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.184078 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.184101 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:06 crc kubenswrapper[4714]: E0129 16:11:06.184254 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:06 crc kubenswrapper[4714]: E0129 16:11:06.184334 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:06 crc kubenswrapper[4714]: E0129 16:11:06.184498 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.189797 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.189829 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.189838 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.189849 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.189859 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:06Z","lastTransitionTime":"2026-01-29T16:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.292088 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.292157 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.292170 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.292192 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.292204 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:06Z","lastTransitionTime":"2026-01-29T16:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.394764 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.394821 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.394841 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.394867 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.394884 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:06Z","lastTransitionTime":"2026-01-29T16:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.498081 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.498138 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.498150 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.498171 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.498185 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:06Z","lastTransitionTime":"2026-01-29T16:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.601802 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.601867 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.601886 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.601913 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.601958 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:06Z","lastTransitionTime":"2026-01-29T16:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.705395 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.705464 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.705475 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.705498 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.705512 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:06Z","lastTransitionTime":"2026-01-29T16:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.808776 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.808820 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.808830 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.808847 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.808859 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:06Z","lastTransitionTime":"2026-01-29T16:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.850368 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 12:30:46.501275408 +0000 UTC Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.911847 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.911918 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.911950 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.911970 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:06 crc kubenswrapper[4714]: I0129 16:11:06.911985 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:06Z","lastTransitionTime":"2026-01-29T16:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.014885 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.014953 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.014965 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.014985 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.014999 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:07Z","lastTransitionTime":"2026-01-29T16:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.117724 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.117806 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.117829 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.117860 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.117883 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:07Z","lastTransitionTime":"2026-01-29T16:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.183415 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:07 crc kubenswrapper[4714]: E0129 16:11:07.183604 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.196387 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.220315 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.220342 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.220352 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.220366 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.220375 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:07Z","lastTransitionTime":"2026-01-29T16:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.322343 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.322571 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.322648 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.322723 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.322784 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:07Z","lastTransitionTime":"2026-01-29T16:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.425284 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.425336 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.425348 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.425365 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.425378 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:07Z","lastTransitionTime":"2026-01-29T16:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.527263 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.527314 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.527324 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.527337 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.527347 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:07Z","lastTransitionTime":"2026-01-29T16:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.630553 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.630620 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.630637 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.630662 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.630682 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:07Z","lastTransitionTime":"2026-01-29T16:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.734363 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.734424 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.734439 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.734462 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.734479 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:07Z","lastTransitionTime":"2026-01-29T16:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.837422 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.837462 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.837471 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.837485 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.837496 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:07Z","lastTransitionTime":"2026-01-29T16:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.850665 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 06:03:55.323231324 +0000 UTC Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.941490 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.941535 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.941546 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.941560 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:07 crc kubenswrapper[4714]: I0129 16:11:07.941571 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:07Z","lastTransitionTime":"2026-01-29T16:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.043880 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.043921 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.043956 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.043972 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.043982 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:08Z","lastTransitionTime":"2026-01-29T16:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.148153 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.148200 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.148214 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.148240 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.148250 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:08Z","lastTransitionTime":"2026-01-29T16:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.183328 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.183380 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:08 crc kubenswrapper[4714]: E0129 16:11:08.183475 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.183325 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:08 crc kubenswrapper[4714]: E0129 16:11:08.183596 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:08 crc kubenswrapper[4714]: E0129 16:11:08.183729 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.251243 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.251337 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.251350 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.251373 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.251389 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:08Z","lastTransitionTime":"2026-01-29T16:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.355275 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.355337 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.355357 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.355382 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.355401 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:08Z","lastTransitionTime":"2026-01-29T16:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.459296 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.459372 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.459427 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.459471 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.459499 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:08Z","lastTransitionTime":"2026-01-29T16:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.563713 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.563777 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.563792 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.563818 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.563861 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:08Z","lastTransitionTime":"2026-01-29T16:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.667693 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.667749 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.667761 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.667781 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.667795 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:08Z","lastTransitionTime":"2026-01-29T16:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.770464 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.770518 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.770531 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.770552 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.770563 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:08Z","lastTransitionTime":"2026-01-29T16:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.850741 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 14:27:48.783025477 +0000 UTC Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.873497 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.873590 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.873609 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.873636 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.873654 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:08Z","lastTransitionTime":"2026-01-29T16:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.977084 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.977157 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.977176 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.977204 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:08 crc kubenswrapper[4714]: I0129 16:11:08.977228 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:08Z","lastTransitionTime":"2026-01-29T16:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.080810 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.080853 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.080864 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.080880 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.080890 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:09Z","lastTransitionTime":"2026-01-29T16:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.183072 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.183109 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.183119 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.183134 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.183145 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:09Z","lastTransitionTime":"2026-01-29T16:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.183196 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:09 crc kubenswrapper[4714]: E0129 16:11:09.183396 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.285420 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.285480 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.285497 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.285520 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.285725 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:09Z","lastTransitionTime":"2026-01-29T16:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.392699 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.392786 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.392799 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.392823 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.392839 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:09Z","lastTransitionTime":"2026-01-29T16:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.496339 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.496656 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.496919 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.497011 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.497100 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:09Z","lastTransitionTime":"2026-01-29T16:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.600186 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.600234 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.600280 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.600303 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.600321 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:09Z","lastTransitionTime":"2026-01-29T16:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.703226 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.703262 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.703271 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.703288 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.703298 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:09Z","lastTransitionTime":"2026-01-29T16:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.807511 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.807551 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.807560 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.807575 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.807584 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:09Z","lastTransitionTime":"2026-01-29T16:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.851348 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 09:24:39.95754176 +0000 UTC Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.910875 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.910917 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.910967 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.910983 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:09 crc kubenswrapper[4714]: I0129 16:11:09.910994 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:09Z","lastTransitionTime":"2026-01-29T16:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.013733 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.013829 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.013849 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.013883 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.013903 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:10Z","lastTransitionTime":"2026-01-29T16:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.117320 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.117439 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.117450 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.117472 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.117484 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:10Z","lastTransitionTime":"2026-01-29T16:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.184180 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.184211 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.184301 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:10 crc kubenswrapper[4714]: E0129 16:11:10.184427 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:10 crc kubenswrapper[4714]: E0129 16:11:10.184582 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:10 crc kubenswrapper[4714]: E0129 16:11:10.184659 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.219979 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.220041 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.220062 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.220090 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.220123 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:10Z","lastTransitionTime":"2026-01-29T16:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.323691 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.323746 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.323758 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.323777 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.323791 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:10Z","lastTransitionTime":"2026-01-29T16:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.425862 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.425920 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.425966 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.425991 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.426008 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:10Z","lastTransitionTime":"2026-01-29T16:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.528739 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.528807 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.528833 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.528865 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.528890 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:10Z","lastTransitionTime":"2026-01-29T16:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.632215 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.632298 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.632323 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.632354 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.632379 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:10Z","lastTransitionTime":"2026-01-29T16:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.736278 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.736324 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.736339 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.736357 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.736366 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:10Z","lastTransitionTime":"2026-01-29T16:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.839575 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.839637 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.839656 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.839680 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.839698 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:10Z","lastTransitionTime":"2026-01-29T16:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.851509 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 00:10:20.282900018 +0000 UTC Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.942819 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.942863 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.942877 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.942893 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:10 crc kubenswrapper[4714]: I0129 16:11:10.942905 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:10Z","lastTransitionTime":"2026-01-29T16:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.045706 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.045742 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.045751 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.045766 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.045776 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:11Z","lastTransitionTime":"2026-01-29T16:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.149165 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.149247 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.149269 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.149299 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.149321 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:11Z","lastTransitionTime":"2026-01-29T16:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.183463 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:11 crc kubenswrapper[4714]: E0129 16:11:11.183661 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.252548 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.252672 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.252691 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.252720 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.252738 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:11Z","lastTransitionTime":"2026-01-29T16:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.355701 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.355751 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.355766 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.355786 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.355801 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:11Z","lastTransitionTime":"2026-01-29T16:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.458891 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.458963 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.458977 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.458996 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.459011 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:11Z","lastTransitionTime":"2026-01-29T16:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.561663 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.561704 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.561713 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.561728 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.561737 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:11Z","lastTransitionTime":"2026-01-29T16:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.664896 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.664965 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.664978 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.664994 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.665005 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:11Z","lastTransitionTime":"2026-01-29T16:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.767936 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.768010 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.768022 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.768041 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.768053 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:11Z","lastTransitionTime":"2026-01-29T16:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.852179 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 17:20:32.040320565 +0000 UTC Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.870911 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.871058 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.871092 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.871123 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.871147 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:11Z","lastTransitionTime":"2026-01-29T16:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.974275 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.974358 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.974384 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.974414 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:11 crc kubenswrapper[4714]: I0129 16:11:11.974437 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:11Z","lastTransitionTime":"2026-01-29T16:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.078171 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.078253 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.078274 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.078302 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.078323 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:12Z","lastTransitionTime":"2026-01-29T16:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.180723 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.180770 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.180780 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.180794 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.180805 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:12Z","lastTransitionTime":"2026-01-29T16:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.183981 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.184063 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:12 crc kubenswrapper[4714]: E0129 16:11:12.184780 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.184803 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:12 crc kubenswrapper[4714]: E0129 16:11:12.185084 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:12 crc kubenswrapper[4714]: E0129 16:11:12.185205 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.283048 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.283433 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.283643 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.283876 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.284141 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:12Z","lastTransitionTime":"2026-01-29T16:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.387318 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.387356 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.387366 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.387382 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.387392 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:12Z","lastTransitionTime":"2026-01-29T16:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.455004 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.455429 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.455637 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.455864 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.456122 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:12Z","lastTransitionTime":"2026-01-29T16:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:12 crc kubenswrapper[4714]: E0129 16:11:12.470031 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:12Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.475083 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.475339 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.475550 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.475765 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.476002 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:12Z","lastTransitionTime":"2026-01-29T16:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:12 crc kubenswrapper[4714]: E0129 16:11:12.498082 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:12Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.503353 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.503409 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.503421 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.503441 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.503454 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:12Z","lastTransitionTime":"2026-01-29T16:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:12 crc kubenswrapper[4714]: E0129 16:11:12.523227 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:12Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.527736 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.527784 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.527795 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.527814 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.527826 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:12Z","lastTransitionTime":"2026-01-29T16:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:12 crc kubenswrapper[4714]: E0129 16:11:12.544892 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:12Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.549100 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.549273 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.549379 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.549483 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.549582 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:12Z","lastTransitionTime":"2026-01-29T16:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:12 crc kubenswrapper[4714]: E0129 16:11:12.567470 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:12Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:12 crc kubenswrapper[4714]: E0129 16:11:12.567628 4714 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.569468 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.569504 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.569516 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.569532 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.569547 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:12Z","lastTransitionTime":"2026-01-29T16:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.671641 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.671700 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.671713 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.671732 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.671746 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:12Z","lastTransitionTime":"2026-01-29T16:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.774501 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.774538 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.774547 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.774561 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.774570 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:12Z","lastTransitionTime":"2026-01-29T16:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.853385 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 05:34:25.726392431 +0000 UTC Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.877222 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.877258 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.877269 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.877283 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.877292 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:12Z","lastTransitionTime":"2026-01-29T16:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.980799 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.980860 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.980878 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.980903 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:12 crc kubenswrapper[4714]: I0129 16:11:12.980920 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:12Z","lastTransitionTime":"2026-01-29T16:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.084024 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.084068 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.084079 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.084095 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.084107 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:13Z","lastTransitionTime":"2026-01-29T16:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.184109 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:13 crc kubenswrapper[4714]: E0129 16:11:13.184283 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.186851 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.186885 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.186895 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.186910 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.186925 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:13Z","lastTransitionTime":"2026-01-29T16:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.290231 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.290302 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.290319 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.290346 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.290366 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:13Z","lastTransitionTime":"2026-01-29T16:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.392792 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.392837 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.392845 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.392860 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.392869 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:13Z","lastTransitionTime":"2026-01-29T16:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.496676 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.496733 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.496747 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.496767 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.496784 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:13Z","lastTransitionTime":"2026-01-29T16:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.601581 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.601623 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.601632 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.601648 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.601657 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:13Z","lastTransitionTime":"2026-01-29T16:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.706263 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.706413 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.706436 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.706499 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.706521 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:13Z","lastTransitionTime":"2026-01-29T16:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.810995 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.811077 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.811097 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.811128 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.811147 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:13Z","lastTransitionTime":"2026-01-29T16:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.853645 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 05:40:46.504149847 +0000 UTC Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.915107 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.915183 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.915202 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.915224 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:13 crc kubenswrapper[4714]: I0129 16:11:13.915237 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:13Z","lastTransitionTime":"2026-01-29T16:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.018380 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.018439 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.018459 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.018481 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.018495 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:14Z","lastTransitionTime":"2026-01-29T16:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.126305 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.126368 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.126379 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.126396 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.126420 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:14Z","lastTransitionTime":"2026-01-29T16:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.184200 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:14 crc kubenswrapper[4714]: E0129 16:11:14.184390 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.184443 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.184635 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:14 crc kubenswrapper[4714]: E0129 16:11:14.184802 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:14 crc kubenswrapper[4714]: E0129 16:11:14.184901 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.186759 4714 scope.go:117] "RemoveContainer" containerID="98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.204871 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2932c3bd-04c7-4494-8d43-03c4524a353f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78109e9e498f3d640934a2c45faf27133d32ece9e35e42ed48ddba720fa7a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc1d7b5f32fe8de4bfd5ba555838989293be538217424c86ea5cedadb8295f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tg8sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.218347 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89403a5-379d-4c3f-a87f-8d2ed63ab368\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0bdc60ea5a5e188b2ed8894bb387084567b294c3a356fed01493d4bd6a7caf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6068173079b1d5cf7de92fe34bd0a4701b11a4dba7ae384f2d0e33f185656107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6565dc4c35c618a239bdc2be6dd45a9057e83573c92c3fc816350eb014c822c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.229120 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.229176 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.229189 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.229214 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.229228 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:14Z","lastTransitionTime":"2026-01-29T16:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.240538 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.257653 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.293800 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:47Z\\\",\\\"message\\\":\\\"ing reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:10:47.314985 6375 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:10:47.315336 6375 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 16:10:47.315637 6375 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0129 16:10:47.315679 6375 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0129 16:10:47.315724 6375 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 16:10:47.315752 6375 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 16:10:47.315774 6375 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0129 16:10:47.315793 6375 factory.go:656] Stopping watch factory\\\\nI0129 16:10:47.315813 6375 ovnkube.go:599] Stopped ovnkube\\\\nI0129 16:10:47.315849 6375 handler.go:208] Removed *v1.Node event handler 2\\\\nI0129 16:10:47.315864 6375 handler.go:208] Removed *v1.Node event handler 7\\\\nI0129 16:10:47.315882 6375 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0129 16:10:47.315889 6375 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-sbnkt_openshift-ovn-kubernetes(04b20f02-6c1e-4082-8233-8f06bda63195)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.309486 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2w92b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791456e8-8d95-4cdb-8fd1-d06a7586b328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2w92b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.327775 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.331844 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.331902 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.331926 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.332064 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.332172 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:14Z","lastTransitionTime":"2026-01-29T16:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.342580 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.358356 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.370583 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.391658 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.407447 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.419303 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.434620 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.434679 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.434698 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.434719 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.434731 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:14Z","lastTransitionTime":"2026-01-29T16:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.436876 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c747ed61a18e27d63630395860ce896242426b1ce46ea5f9d00534b808804a58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:11:03Z\\\",\\\"message\\\":\\\"2026-01-29T16:10:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_27c11876-c5ae-4d35-b720-78ab09bfac0c\\\\n2026-01-29T16:10:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_27c11876-c5ae-4d35-b720-78ab09bfac0c to /host/opt/cni/bin/\\\\n2026-01-29T16:10:18Z [verbose] multus-daemon started\\\\n2026-01-29T16:10:18Z [verbose] Readiness Indicator file check\\\\n2026-01-29T16:11:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.451313 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.469629 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a65c2b0-9568-4a06-8073-93ec194b4ee0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cb2f8d441042c4c95e3cc056f991565c18bd93dcb0d61f3735e2451ff439a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86530d623401c3cf5ebe44fcc1c2110a5f4dde059a5d27f74e118d5ca1df40dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86530d623401c3cf5ebe44fcc1c2110a5f4dde059a5d27f74e118d5ca1df40dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.492361 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.508810 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.529383 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.538128 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.538195 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.538217 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.538240 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.538253 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:14Z","lastTransitionTime":"2026-01-29T16:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.640834 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.640892 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.640910 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.640958 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.640975 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:14Z","lastTransitionTime":"2026-01-29T16:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.689691 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-sbnkt_04b20f02-6c1e-4082-8233-8f06bda63195/ovnkube-controller/2.log" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.692874 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerStarted","Data":"95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365"} Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.694363 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.713738 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.775672 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.775722 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.775732 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.775748 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.775757 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:14Z","lastTransitionTime":"2026-01-29T16:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.781732 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.800245 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.821172 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a65c2b0-9568-4a06-8073-93ec194b4ee0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cb2f8d441042c4c95e3cc056f991565c18bd93dcb0d61f3735e2451ff439a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86530d623401c3cf5ebe44fcc1c2110a5f4dde059a5d27f74e118d5ca1df40dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86530d623401c3cf5ebe44fcc1c2110a5f4dde059a5d27f74e118d5ca1df40dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.855056 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 09:01:16.247836598 +0000 UTC Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.856566 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.869212 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.878235 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.878266 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.878278 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.878316 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.878331 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:14Z","lastTransitionTime":"2026-01-29T16:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.894011 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:47Z\\\",\\\"message\\\":\\\"ing reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:10:47.314985 6375 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:10:47.315336 6375 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 16:10:47.315637 6375 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0129 16:10:47.315679 6375 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0129 16:10:47.315724 6375 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 16:10:47.315752 6375 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 16:10:47.315774 6375 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0129 16:10:47.315793 6375 factory.go:656] Stopping watch factory\\\\nI0129 16:10:47.315813 6375 ovnkube.go:599] Stopped ovnkube\\\\nI0129 16:10:47.315849 6375 handler.go:208] Removed *v1.Node event handler 2\\\\nI0129 16:10:47.315864 6375 handler.go:208] Removed *v1.Node event handler 7\\\\nI0129 16:10:47.315882 6375 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0129 16:10:47.315889 6375 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:11:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.906681 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2932c3bd-04c7-4494-8d43-03c4524a353f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78109e9e498f3d640934a2c45faf27133d32ece9e35e42ed48ddba720fa7a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc1d7b5f32fe8de4bfd5ba555838989293be538217424c86ea5cedadb8295f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tg8sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.918738 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89403a5-379d-4c3f-a87f-8d2ed63ab368\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0bdc60ea5a5e188b2ed8894bb387084567b294c3a356fed01493d4bd6a7caf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6068173079b1d5cf7de92fe34bd0a4701b11a4dba7ae384f2d0e33f185656107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6565dc4c35c618a239bdc2be6dd45a9057e83573c92c3fc816350eb014c822c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.929724 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.944969 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.955910 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.967109 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2w92b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791456e8-8d95-4cdb-8fd1-d06a7586b328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2w92b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.980527 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.980573 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.980584 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.980604 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.980615 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:14Z","lastTransitionTime":"2026-01-29T16:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:14 crc kubenswrapper[4714]: I0129 16:11:14.992899 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:14Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.009346 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.025228 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.039339 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c747ed61a18e27d63630395860ce896242426b1ce46ea5f9d00534b808804a58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:11:03Z\\\",\\\"message\\\":\\\"2026-01-29T16:10:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_27c11876-c5ae-4d35-b720-78ab09bfac0c\\\\n2026-01-29T16:10:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_27c11876-c5ae-4d35-b720-78ab09bfac0c to /host/opt/cni/bin/\\\\n2026-01-29T16:10:18Z [verbose] multus-daemon started\\\\n2026-01-29T16:10:18Z [verbose] Readiness Indicator file check\\\\n2026-01-29T16:11:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.057976 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.078545 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:15Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.083586 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.083653 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.083675 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.083701 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.083718 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:15Z","lastTransitionTime":"2026-01-29T16:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.184079 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:15 crc kubenswrapper[4714]: E0129 16:11:15.185116 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.186045 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.186092 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.186104 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.186120 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.186133 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:15Z","lastTransitionTime":"2026-01-29T16:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.288732 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.288785 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.288800 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.288819 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.288831 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:15Z","lastTransitionTime":"2026-01-29T16:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.392486 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.393027 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.393050 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.393071 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.393085 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:15Z","lastTransitionTime":"2026-01-29T16:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.495789 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.495856 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.495870 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.495894 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.495911 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:15Z","lastTransitionTime":"2026-01-29T16:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.599404 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.599465 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.599480 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.599520 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.599537 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:15Z","lastTransitionTime":"2026-01-29T16:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.701130 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.701170 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.701182 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.701198 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.701209 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:15Z","lastTransitionTime":"2026-01-29T16:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.805428 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.805493 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.805512 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.805536 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.805552 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:15Z","lastTransitionTime":"2026-01-29T16:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.855994 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 01:39:29.579521206 +0000 UTC Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.909197 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.909238 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.909251 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.909268 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:15 crc kubenswrapper[4714]: I0129 16:11:15.909279 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:15Z","lastTransitionTime":"2026-01-29T16:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.012860 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.012970 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.012989 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.013014 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.013032 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:16Z","lastTransitionTime":"2026-01-29T16:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.116350 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.116424 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.116447 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.116472 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.116493 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:16Z","lastTransitionTime":"2026-01-29T16:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.183670 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.183710 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.183779 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:16 crc kubenswrapper[4714]: E0129 16:11:16.183835 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:16 crc kubenswrapper[4714]: E0129 16:11:16.183973 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:16 crc kubenswrapper[4714]: E0129 16:11:16.190227 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.223731 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.223791 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.223809 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.223835 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.223901 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:16Z","lastTransitionTime":"2026-01-29T16:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.332365 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.332703 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.332718 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.332738 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.332751 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:16Z","lastTransitionTime":"2026-01-29T16:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.435220 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.435256 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.435268 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.435284 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.435296 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:16Z","lastTransitionTime":"2026-01-29T16:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.538443 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.538536 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.538555 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.538584 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.538603 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:16Z","lastTransitionTime":"2026-01-29T16:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.641488 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.641533 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.641541 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.641556 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.641565 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:16Z","lastTransitionTime":"2026-01-29T16:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.703556 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-sbnkt_04b20f02-6c1e-4082-8233-8f06bda63195/ovnkube-controller/3.log" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.704477 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-sbnkt_04b20f02-6c1e-4082-8233-8f06bda63195/ovnkube-controller/2.log" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.708899 4714 generic.go:334] "Generic (PLEG): container finished" podID="04b20f02-6c1e-4082-8233-8f06bda63195" containerID="95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365" exitCode=1 Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.708993 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerDied","Data":"95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365"} Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.709074 4714 scope.go:117] "RemoveContainer" containerID="98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.710088 4714 scope.go:117] "RemoveContainer" containerID="95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365" Jan 29 16:11:16 crc kubenswrapper[4714]: E0129 16:11:16.710417 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-sbnkt_openshift-ovn-kubernetes(04b20f02-6c1e-4082-8233-8f06bda63195)\"" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.731924 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.745146 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.745214 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.745235 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.745264 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.745288 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:16Z","lastTransitionTime":"2026-01-29T16:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.768466 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:47Z\\\",\\\"message\\\":\\\"ing reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:10:47.314985 6375 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:10:47.315336 6375 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 16:10:47.315637 6375 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0129 16:10:47.315679 6375 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0129 16:10:47.315724 6375 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 16:10:47.315752 6375 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 16:10:47.315774 6375 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0129 16:10:47.315793 6375 factory.go:656] Stopping watch factory\\\\nI0129 16:10:47.315813 6375 ovnkube.go:599] Stopped ovnkube\\\\nI0129 16:10:47.315849 6375 handler.go:208] Removed *v1.Node event handler 2\\\\nI0129 16:10:47.315864 6375 handler.go:208] Removed *v1.Node event handler 7\\\\nI0129 16:10:47.315882 6375 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0129 16:10:47.315889 6375 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:11:16Z\\\",\\\"message\\\":\\\"] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:11:15.260241 6760 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:11:15.260645 6760 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0129 16:11:15.261304 6760 obj_retry.go:551] Creating *factory.egressNode crc took: 14.368203ms\\\\nI0129 16:11:15.261337 6760 factory.go:1336] Added *v1.Node event handler 7\\\\nI0129 16:11:15.261379 6760 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0129 16:11:15.261465 6760 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0129 16:11:15.261482 6760 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0129 16:11:15.261500 6760 factory.go:656] Stopping watch factory\\\\nI0129 16:11:15.261522 6760 handler.go:208] Removed *v1.Node event handler 7\\\\nI0129 16:11:15.261535 6760 handler.go:208] Removed *v1.Node event handler 2\\\\nI0129 16:11:15.261674 6760 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0129 16:11:15.261875 6760 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0129 16:11:15.261909 6760 ovnkube.go:599] Stopped ovnkube\\\\nI0129 16:11:15.261936 6760 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0129 16:11:15.262026 6760 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:11:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.788270 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2932c3bd-04c7-4494-8d43-03c4524a353f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78109e9e498f3d640934a2c45faf27133d32ece9e35e42ed48ddba720fa7a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc1d7b5f32fe8de4bfd5ba555838989293be538217424c86ea5cedadb8295f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tg8sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.808486 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89403a5-379d-4c3f-a87f-8d2ed63ab368\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0bdc60ea5a5e188b2ed8894bb387084567b294c3a356fed01493d4bd6a7caf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6068173079b1d5cf7de92fe34bd0a4701b11a4dba7ae384f2d0e33f185656107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6565dc4c35c618a239bdc2be6dd45a9057e83573c92c3fc816350eb014c822c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.830821 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.848091 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.848184 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.848204 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.848680 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.848738 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:16Z","lastTransitionTime":"2026-01-29T16:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.856769 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 15:45:18.064304185 +0000 UTC Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.858406 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.877177 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.896727 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2w92b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791456e8-8d95-4cdb-8fd1-d06a7586b328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2w92b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.920932 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.942987 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.953002 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.953056 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.953071 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.953093 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.953110 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:16Z","lastTransitionTime":"2026-01-29T16:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.961241 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:16 crc kubenswrapper[4714]: I0129 16:11:16.984706 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c747ed61a18e27d63630395860ce896242426b1ce46ea5f9d00534b808804a58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:11:03Z\\\",\\\"message\\\":\\\"2026-01-29T16:10:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_27c11876-c5ae-4d35-b720-78ab09bfac0c\\\\n2026-01-29T16:10:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_27c11876-c5ae-4d35-b720-78ab09bfac0c to /host/opt/cni/bin/\\\\n2026-01-29T16:10:18Z [verbose] multus-daemon started\\\\n2026-01-29T16:10:18Z [verbose] Readiness Indicator file check\\\\n2026-01-29T16:11:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:16Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.010465 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.032027 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.056171 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.057476 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.057522 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.057539 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.057564 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.057584 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:17Z","lastTransitionTime":"2026-01-29T16:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.081412 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.095123 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.110387 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a65c2b0-9568-4a06-8073-93ec194b4ee0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cb2f8d441042c4c95e3cc056f991565c18bd93dcb0d61f3735e2451ff439a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86530d623401c3cf5ebe44fcc1c2110a5f4dde059a5d27f74e118d5ca1df40dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86530d623401c3cf5ebe44fcc1c2110a5f4dde059a5d27f74e118d5ca1df40dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.140829 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:17Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.160078 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.160138 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.160157 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.160182 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.160200 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:17Z","lastTransitionTime":"2026-01-29T16:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.183189 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:17 crc kubenswrapper[4714]: E0129 16:11:17.183367 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.263970 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.264054 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.264074 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.264100 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.264118 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:17Z","lastTransitionTime":"2026-01-29T16:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.367346 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.367410 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.367429 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.367454 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.367473 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:17Z","lastTransitionTime":"2026-01-29T16:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.470690 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.470785 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.470814 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.470846 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.470873 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:17Z","lastTransitionTime":"2026-01-29T16:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.573590 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.573673 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.573690 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.573714 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.573730 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:17Z","lastTransitionTime":"2026-01-29T16:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.676317 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.676355 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.676365 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.676382 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.676391 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:17Z","lastTransitionTime":"2026-01-29T16:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.779840 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.779904 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.779921 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.779983 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.780002 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:17Z","lastTransitionTime":"2026-01-29T16:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.857735 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 21:07:37.106953819 +0000 UTC Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.883357 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.883430 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.883441 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.883456 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.883466 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:17Z","lastTransitionTime":"2026-01-29T16:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.986323 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.986419 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.986442 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.986477 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:17 crc kubenswrapper[4714]: I0129 16:11:17.986504 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:17Z","lastTransitionTime":"2026-01-29T16:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.090513 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.090598 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.090620 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.090658 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.090683 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:18Z","lastTransitionTime":"2026-01-29T16:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.183651 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.183751 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:18 crc kubenswrapper[4714]: E0129 16:11:18.183830 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.183861 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:18 crc kubenswrapper[4714]: E0129 16:11:18.183998 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:18 crc kubenswrapper[4714]: E0129 16:11:18.184234 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.193499 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.193574 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.193597 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.193626 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.193647 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:18Z","lastTransitionTime":"2026-01-29T16:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.296301 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.296397 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.296425 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.296459 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.296483 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:18Z","lastTransitionTime":"2026-01-29T16:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.315755 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:11:18 crc kubenswrapper[4714]: E0129 16:11:18.316118 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:22.31608611 +0000 UTC m=+148.836587230 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.400374 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.400441 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.400461 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.400490 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.400507 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:18Z","lastTransitionTime":"2026-01-29T16:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.417410 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.417479 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.417516 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:18 crc kubenswrapper[4714]: E0129 16:11:18.417549 4714 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.417570 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:18 crc kubenswrapper[4714]: E0129 16:11:18.417639 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 16:12:22.417612243 +0000 UTC m=+148.938113403 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 16:11:18 crc kubenswrapper[4714]: E0129 16:11:18.417795 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 16:11:18 crc kubenswrapper[4714]: E0129 16:11:18.417835 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 16:11:18 crc kubenswrapper[4714]: E0129 16:11:18.417840 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 16:11:18 crc kubenswrapper[4714]: E0129 16:11:18.417862 4714 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 16:11:18 crc kubenswrapper[4714]: E0129 16:11:18.417870 4714 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:11:18 crc kubenswrapper[4714]: E0129 16:11:18.417874 4714 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 16:11:18 crc kubenswrapper[4714]: E0129 16:11:18.417983 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 16:12:22.417923592 +0000 UTC m=+148.938424752 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:11:18 crc kubenswrapper[4714]: E0129 16:11:18.417882 4714 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:11:18 crc kubenswrapper[4714]: E0129 16:11:18.418014 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 16:12:22.418001944 +0000 UTC m=+148.938503094 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 16:11:18 crc kubenswrapper[4714]: E0129 16:11:18.418062 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 16:12:22.418035365 +0000 UTC m=+148.938536515 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.504244 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.504313 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.504330 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.504357 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.504376 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:18Z","lastTransitionTime":"2026-01-29T16:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.607660 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.607845 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.607870 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.607916 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.607978 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:18Z","lastTransitionTime":"2026-01-29T16:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.712109 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.712175 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.712193 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.712218 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.712236 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:18Z","lastTransitionTime":"2026-01-29T16:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.716218 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-sbnkt_04b20f02-6c1e-4082-8233-8f06bda63195/ovnkube-controller/3.log" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.826496 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.826543 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.826554 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.826572 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.826586 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:18Z","lastTransitionTime":"2026-01-29T16:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.857887 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 03:28:11.911023231 +0000 UTC Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.929640 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.929712 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.929736 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.929766 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:18 crc kubenswrapper[4714]: I0129 16:11:18.929787 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:18Z","lastTransitionTime":"2026-01-29T16:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.033219 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.033277 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.033294 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.033318 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.033336 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:19Z","lastTransitionTime":"2026-01-29T16:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.136771 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.136840 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.136857 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.136887 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.136910 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:19Z","lastTransitionTime":"2026-01-29T16:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.183633 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:19 crc kubenswrapper[4714]: E0129 16:11:19.183928 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.239638 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.239691 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.239710 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.239734 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.239792 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:19Z","lastTransitionTime":"2026-01-29T16:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.342916 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.343035 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.343056 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.343085 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.343108 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:19Z","lastTransitionTime":"2026-01-29T16:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.446208 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.446249 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.446296 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.446314 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.446327 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:19Z","lastTransitionTime":"2026-01-29T16:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.549374 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.549451 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.549474 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.549504 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.549527 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:19Z","lastTransitionTime":"2026-01-29T16:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.652793 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.653253 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.653271 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.653296 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.653313 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:19Z","lastTransitionTime":"2026-01-29T16:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.755976 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.756056 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.756083 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.756118 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.756141 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:19Z","lastTransitionTime":"2026-01-29T16:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.858010 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 08:26:03.331778385 +0000 UTC Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.859262 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.859313 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.859335 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.859365 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.859382 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:19Z","lastTransitionTime":"2026-01-29T16:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.966428 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.966517 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.966543 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.966576 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:19 crc kubenswrapper[4714]: I0129 16:11:19.966609 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:19Z","lastTransitionTime":"2026-01-29T16:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.070418 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.070494 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.070531 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.070562 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.070583 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:20Z","lastTransitionTime":"2026-01-29T16:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.174065 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.174148 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.174187 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.174220 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.174242 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:20Z","lastTransitionTime":"2026-01-29T16:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.183323 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.183363 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:20 crc kubenswrapper[4714]: E0129 16:11:20.183544 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.183629 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:20 crc kubenswrapper[4714]: E0129 16:11:20.183814 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:20 crc kubenswrapper[4714]: E0129 16:11:20.184004 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.276747 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.276824 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.276842 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.276867 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.276887 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:20Z","lastTransitionTime":"2026-01-29T16:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.379975 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.380040 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.380053 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.380073 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.380088 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:20Z","lastTransitionTime":"2026-01-29T16:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.483346 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.483416 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.483435 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.483461 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.483484 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:20Z","lastTransitionTime":"2026-01-29T16:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.586589 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.586632 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.586643 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.586660 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.586672 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:20Z","lastTransitionTime":"2026-01-29T16:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.690154 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.690211 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.690231 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.690254 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.690271 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:20Z","lastTransitionTime":"2026-01-29T16:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.793075 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.793138 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.793162 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.793192 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.793215 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:20Z","lastTransitionTime":"2026-01-29T16:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.858203 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 07:17:38.887233849 +0000 UTC Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.896644 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.896717 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.896741 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.896770 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:20 crc kubenswrapper[4714]: I0129 16:11:20.896792 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:20Z","lastTransitionTime":"2026-01-29T16:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.000345 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.000414 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.000435 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.000462 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.000484 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:21Z","lastTransitionTime":"2026-01-29T16:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.103743 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.103810 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.103829 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.103850 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.103865 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:21Z","lastTransitionTime":"2026-01-29T16:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.183214 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:21 crc kubenswrapper[4714]: E0129 16:11:21.183344 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.205902 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.205960 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.205979 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.205998 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.206009 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:21Z","lastTransitionTime":"2026-01-29T16:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.309091 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.309144 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.309157 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.309175 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.309187 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:21Z","lastTransitionTime":"2026-01-29T16:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.412440 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.412487 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.412500 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.412517 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.412528 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:21Z","lastTransitionTime":"2026-01-29T16:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.515769 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.515821 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.515834 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.515855 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.515868 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:21Z","lastTransitionTime":"2026-01-29T16:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.617924 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.617995 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.618007 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.618024 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.618038 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:21Z","lastTransitionTime":"2026-01-29T16:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.721224 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.721280 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.721295 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.721317 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.721333 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:21Z","lastTransitionTime":"2026-01-29T16:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.824326 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.824405 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.824430 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.824456 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.824477 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:21Z","lastTransitionTime":"2026-01-29T16:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.858339 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 14:29:59.858478558 +0000 UTC Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.927057 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.927127 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.927146 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.927173 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:21 crc kubenswrapper[4714]: I0129 16:11:21.927194 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:21Z","lastTransitionTime":"2026-01-29T16:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.030478 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.030552 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.030570 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.030595 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.030615 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:22Z","lastTransitionTime":"2026-01-29T16:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.136526 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.136578 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.136588 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.136609 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.136619 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:22Z","lastTransitionTime":"2026-01-29T16:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.184107 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.184202 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:22 crc kubenswrapper[4714]: E0129 16:11:22.184238 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.184334 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:22 crc kubenswrapper[4714]: E0129 16:11:22.184393 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:22 crc kubenswrapper[4714]: E0129 16:11:22.184472 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.238625 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.238652 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.238660 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.238670 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.238678 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:22Z","lastTransitionTime":"2026-01-29T16:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.341239 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.341310 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.341334 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.341361 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.341378 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:22Z","lastTransitionTime":"2026-01-29T16:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.444539 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.444587 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.444603 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.444625 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.444641 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:22Z","lastTransitionTime":"2026-01-29T16:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.547519 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.547592 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.547619 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.547650 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.547674 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:22Z","lastTransitionTime":"2026-01-29T16:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.650919 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.650997 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.651019 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.651046 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.651063 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:22Z","lastTransitionTime":"2026-01-29T16:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.712471 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.712517 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.712533 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.712550 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.712564 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:22Z","lastTransitionTime":"2026-01-29T16:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:22 crc kubenswrapper[4714]: E0129 16:11:22.729605 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:22Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.733891 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.733953 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.733965 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.733980 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.733989 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:22Z","lastTransitionTime":"2026-01-29T16:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:22 crc kubenswrapper[4714]: E0129 16:11:22.745813 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:22Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.749363 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.749435 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.749460 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.749493 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.749518 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:22Z","lastTransitionTime":"2026-01-29T16:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:22 crc kubenswrapper[4714]: E0129 16:11:22.770543 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:22Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.775742 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.775810 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.775833 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.775862 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.775886 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:22Z","lastTransitionTime":"2026-01-29T16:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:22 crc kubenswrapper[4714]: E0129 16:11:22.796625 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:22Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.801598 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.801885 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.802092 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.802253 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.802420 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:22Z","lastTransitionTime":"2026-01-29T16:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:22 crc kubenswrapper[4714]: E0129 16:11:22.817844 4714 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"856e4040-197b-4e74-9239-c0ebcf6976ae\\\",\\\"systemUUID\\\":\\\"1ab8f43b-7f84-4fd2-a80a-2aae14146bf4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:22Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:22 crc kubenswrapper[4714]: E0129 16:11:22.818005 4714 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.819888 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.819952 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.819974 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.819995 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.820008 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:22Z","lastTransitionTime":"2026-01-29T16:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.859395 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 08:19:51.843447223 +0000 UTC Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.922590 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.922668 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.922693 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.922722 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:22 crc kubenswrapper[4714]: I0129 16:11:22.922745 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:22Z","lastTransitionTime":"2026-01-29T16:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.025829 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.025901 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.025913 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.025959 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.025971 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:23Z","lastTransitionTime":"2026-01-29T16:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.127692 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.127751 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.127764 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.127782 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.127798 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:23Z","lastTransitionTime":"2026-01-29T16:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.183500 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:23 crc kubenswrapper[4714]: E0129 16:11:23.183682 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.230614 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.230699 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.230728 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.230760 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.230782 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:23Z","lastTransitionTime":"2026-01-29T16:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.333709 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.333779 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.333798 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.333823 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.333840 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:23Z","lastTransitionTime":"2026-01-29T16:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.437729 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.437793 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.437812 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.437836 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.437852 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:23Z","lastTransitionTime":"2026-01-29T16:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.540755 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.540841 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.540867 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.540897 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.540919 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:23Z","lastTransitionTime":"2026-01-29T16:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.646449 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.646508 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.646530 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.646555 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.646574 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:23Z","lastTransitionTime":"2026-01-29T16:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.749140 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.749203 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.749225 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.749255 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.749275 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:23Z","lastTransitionTime":"2026-01-29T16:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.851757 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.851822 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.851842 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.851868 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.851887 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:23Z","lastTransitionTime":"2026-01-29T16:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.860286 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 23:56:50.673855057 +0000 UTC Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.955653 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.955723 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.955744 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.955772 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:23 crc kubenswrapper[4714]: I0129 16:11:23.955793 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:23Z","lastTransitionTime":"2026-01-29T16:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.058737 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.058792 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.058804 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.058821 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.058833 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:24Z","lastTransitionTime":"2026-01-29T16:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.161679 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.161742 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.161759 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.161785 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.161802 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:24Z","lastTransitionTime":"2026-01-29T16:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.184073 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.184162 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.184224 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:24 crc kubenswrapper[4714]: E0129 16:11:24.184367 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:24 crc kubenswrapper[4714]: E0129 16:11:24.184666 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:24 crc kubenswrapper[4714]: E0129 16:11:24.184806 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.204027 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.221525 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a65c2b0-9568-4a06-8073-93ec194b4ee0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cb2f8d441042c4c95e3cc056f991565c18bd93dcb0d61f3735e2451ff439a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86530d623401c3cf5ebe44fcc1c2110a5f4dde059a5d27f74e118d5ca1df40dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86530d623401c3cf5ebe44fcc1c2110a5f4dde059a5d27f74e118d5ca1df40dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.249838 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.264969 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.265011 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.265028 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.265047 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.265074 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:24Z","lastTransitionTime":"2026-01-29T16:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.268927 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.294240 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.310843 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2932c3bd-04c7-4494-8d43-03c4524a353f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78109e9e498f3d640934a2c45faf27133d32ece9e35e42ed48ddba720fa7a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc1d7b5f32fe8de4bfd5ba555838989293be538217424c86ea5cedadb8295f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tg8sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.325685 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89403a5-379d-4c3f-a87f-8d2ed63ab368\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0bdc60ea5a5e188b2ed8894bb387084567b294c3a356fed01493d4bd6a7caf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6068173079b1d5cf7de92fe34bd0a4701b11a4dba7ae384f2d0e33f185656107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6565dc4c35c618a239bdc2be6dd45a9057e83573c92c3fc816350eb014c822c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.340183 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.350328 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.367053 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.367098 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.367111 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.367129 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.367141 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:24Z","lastTransitionTime":"2026-01-29T16:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.371067 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d5d7851e659633c2e0da4975fa0e9e36d8d1f1b9d725661500f818f6268bcc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:10:47Z\\\",\\\"message\\\":\\\"ing reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:10:47.314985 6375 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:10:47.315336 6375 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 16:10:47.315637 6375 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0129 16:10:47.315679 6375 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0129 16:10:47.315724 6375 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 16:10:47.315752 6375 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 16:10:47.315774 6375 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0129 16:10:47.315793 6375 factory.go:656] Stopping watch factory\\\\nI0129 16:10:47.315813 6375 ovnkube.go:599] Stopped ovnkube\\\\nI0129 16:10:47.315849 6375 handler.go:208] Removed *v1.Node event handler 2\\\\nI0129 16:10:47.315864 6375 handler.go:208] Removed *v1.Node event handler 7\\\\nI0129 16:10:47.315882 6375 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0129 16:10:47.315889 6375 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:11:16Z\\\",\\\"message\\\":\\\"] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:11:15.260241 6760 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:11:15.260645 6760 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0129 16:11:15.261304 6760 obj_retry.go:551] Creating *factory.egressNode crc took: 14.368203ms\\\\nI0129 16:11:15.261337 6760 factory.go:1336] Added *v1.Node event handler 7\\\\nI0129 16:11:15.261379 6760 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0129 16:11:15.261465 6760 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0129 16:11:15.261482 6760 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0129 16:11:15.261500 6760 factory.go:656] Stopping watch factory\\\\nI0129 16:11:15.261522 6760 handler.go:208] Removed *v1.Node event handler 7\\\\nI0129 16:11:15.261535 6760 handler.go:208] Removed *v1.Node event handler 2\\\\nI0129 16:11:15.261674 6760 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0129 16:11:15.261875 6760 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0129 16:11:15.261909 6760 ovnkube.go:599] Stopped ovnkube\\\\nI0129 16:11:15.261936 6760 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0129 16:11:15.262026 6760 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:11:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.381834 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2w92b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791456e8-8d95-4cdb-8fd1-d06a7586b328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2w92b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.392984 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.405037 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.419768 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.429985 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.444227 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.459575 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.467910 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.468791 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.468813 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.468823 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.468838 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.468849 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:24Z","lastTransitionTime":"2026-01-29T16:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.477570 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c747ed61a18e27d63630395860ce896242426b1ce46ea5f9d00534b808804a58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:11:03Z\\\",\\\"message\\\":\\\"2026-01-29T16:10:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_27c11876-c5ae-4d35-b720-78ab09bfac0c\\\\n2026-01-29T16:10:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_27c11876-c5ae-4d35-b720-78ab09bfac0c to /host/opt/cni/bin/\\\\n2026-01-29T16:10:18Z [verbose] multus-daemon started\\\\n2026-01-29T16:10:18Z [verbose] Readiness Indicator file check\\\\n2026-01-29T16:11:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:24Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.570847 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.570916 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.570967 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.570996 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.571017 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:24Z","lastTransitionTime":"2026-01-29T16:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.673249 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.673320 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.673341 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.673401 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.673423 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:24Z","lastTransitionTime":"2026-01-29T16:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.776130 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.776171 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.776181 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.776196 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.776208 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:24Z","lastTransitionTime":"2026-01-29T16:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.860618 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 18:42:32.7015951 +0000 UTC Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.880027 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.880082 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.880106 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.880377 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.880442 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:24Z","lastTransitionTime":"2026-01-29T16:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.983678 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.983753 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.983772 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.983795 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:24 crc kubenswrapper[4714]: I0129 16:11:24.983812 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:24Z","lastTransitionTime":"2026-01-29T16:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.087121 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.087184 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.087210 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.087239 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.087261 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:25Z","lastTransitionTime":"2026-01-29T16:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.184030 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:25 crc kubenswrapper[4714]: E0129 16:11:25.184227 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.190756 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.190818 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.190837 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.190859 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.190875 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:25Z","lastTransitionTime":"2026-01-29T16:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.294254 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.294329 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.294350 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.294381 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.294404 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:25Z","lastTransitionTime":"2026-01-29T16:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.396440 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.396501 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.396511 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.396524 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.396533 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:25Z","lastTransitionTime":"2026-01-29T16:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.500340 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.500379 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.500389 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.500404 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.500414 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:25Z","lastTransitionTime":"2026-01-29T16:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.603123 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.603269 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.603295 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.603374 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.603454 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:25Z","lastTransitionTime":"2026-01-29T16:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.706666 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.706710 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.706719 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.706735 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.706744 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:25Z","lastTransitionTime":"2026-01-29T16:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.809163 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.809254 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.809267 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.809285 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.809297 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:25Z","lastTransitionTime":"2026-01-29T16:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.861601 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 12:21:39.719320224 +0000 UTC Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.917379 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.917446 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.917471 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.917500 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:25 crc kubenswrapper[4714]: I0129 16:11:25.917521 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:25Z","lastTransitionTime":"2026-01-29T16:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.019893 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.019978 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.019997 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.020021 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.020038 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:26Z","lastTransitionTime":"2026-01-29T16:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.123373 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.123435 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.123463 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.123486 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.123501 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:26Z","lastTransitionTime":"2026-01-29T16:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.184258 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.184315 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.184388 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:26 crc kubenswrapper[4714]: E0129 16:11:26.184471 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:26 crc kubenswrapper[4714]: E0129 16:11:26.184628 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:26 crc kubenswrapper[4714]: E0129 16:11:26.185065 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.225973 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.226045 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.226070 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.226105 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.226128 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:26Z","lastTransitionTime":"2026-01-29T16:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.329763 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.329851 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.329878 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.329912 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.329969 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:26Z","lastTransitionTime":"2026-01-29T16:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.432163 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.432239 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.432263 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.432295 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.432317 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:26Z","lastTransitionTime":"2026-01-29T16:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.534954 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.535030 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.535047 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.535068 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.535083 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:26Z","lastTransitionTime":"2026-01-29T16:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.638500 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.638554 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.638567 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.638586 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.638599 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:26Z","lastTransitionTime":"2026-01-29T16:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.741857 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.741919 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.741988 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.742015 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.742033 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:26Z","lastTransitionTime":"2026-01-29T16:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.845348 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.845432 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.845459 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.845490 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.845512 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:26Z","lastTransitionTime":"2026-01-29T16:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.862692 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 14:24:11.552722654 +0000 UTC Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.947780 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.947832 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.947849 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.947874 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:26 crc kubenswrapper[4714]: I0129 16:11:26.947891 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:26Z","lastTransitionTime":"2026-01-29T16:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.051283 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.051319 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.051329 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.051344 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.051354 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:27Z","lastTransitionTime":"2026-01-29T16:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.154269 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.154333 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.154351 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.154374 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.154392 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:27Z","lastTransitionTime":"2026-01-29T16:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.184191 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:27 crc kubenswrapper[4714]: E0129 16:11:27.184459 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.257633 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.257690 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.257706 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.257728 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.258021 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:27Z","lastTransitionTime":"2026-01-29T16:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.360748 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.360833 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.360855 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.360885 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.360926 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:27Z","lastTransitionTime":"2026-01-29T16:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.464637 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.464708 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.464732 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.464760 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.464783 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:27Z","lastTransitionTime":"2026-01-29T16:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.568362 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.568427 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.568444 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.568469 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.568488 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:27Z","lastTransitionTime":"2026-01-29T16:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.671842 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.671907 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.671926 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.671992 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.672010 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:27Z","lastTransitionTime":"2026-01-29T16:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.774674 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.774749 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.774773 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.774804 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.774829 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:27Z","lastTransitionTime":"2026-01-29T16:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.862912 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 23:58:49.767703108 +0000 UTC Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.878898 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.878993 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.879013 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.879036 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.879050 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:27Z","lastTransitionTime":"2026-01-29T16:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.982196 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.982297 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.982319 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.982345 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:27 crc kubenswrapper[4714]: I0129 16:11:27.982362 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:27Z","lastTransitionTime":"2026-01-29T16:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.085790 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.085855 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.085873 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.085900 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.085918 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:28Z","lastTransitionTime":"2026-01-29T16:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.184006 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.184119 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.184035 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:28 crc kubenswrapper[4714]: E0129 16:11:28.184270 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:28 crc kubenswrapper[4714]: E0129 16:11:28.184267 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:28 crc kubenswrapper[4714]: E0129 16:11:28.184473 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.188080 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.188188 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.188257 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.188296 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.188364 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:28Z","lastTransitionTime":"2026-01-29T16:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.291708 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.291778 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.291802 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.291827 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.291847 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:28Z","lastTransitionTime":"2026-01-29T16:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.394861 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.394910 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.394922 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.394961 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.394974 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:28Z","lastTransitionTime":"2026-01-29T16:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.497570 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.497622 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.497652 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.497672 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.497682 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:28Z","lastTransitionTime":"2026-01-29T16:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.601200 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.601296 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.601320 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.601353 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.601377 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:28Z","lastTransitionTime":"2026-01-29T16:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.705139 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.705210 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.705230 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.705255 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.705273 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:28Z","lastTransitionTime":"2026-01-29T16:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.807673 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.807729 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.807752 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.807779 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.807800 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:28Z","lastTransitionTime":"2026-01-29T16:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.863464 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 05:26:17.802506184 +0000 UTC Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.910628 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.910790 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.910819 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.910895 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:28 crc kubenswrapper[4714]: I0129 16:11:28.910921 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:28Z","lastTransitionTime":"2026-01-29T16:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.014336 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.014408 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.014431 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.014462 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.014486 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:29Z","lastTransitionTime":"2026-01-29T16:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.117635 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.117700 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.117716 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.117740 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.117759 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:29Z","lastTransitionTime":"2026-01-29T16:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.184155 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:29 crc kubenswrapper[4714]: E0129 16:11:29.184372 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.185106 4714 scope.go:117] "RemoveContainer" containerID="95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365" Jan 29 16:11:29 crc kubenswrapper[4714]: E0129 16:11:29.185299 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-sbnkt_openshift-ovn-kubernetes(04b20f02-6c1e-4082-8233-8f06bda63195)\"" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.205452 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89403a5-379d-4c3f-a87f-8d2ed63ab368\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0bdc60ea5a5e188b2ed8894bb387084567b294c3a356fed01493d4bd6a7caf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6068173079b1d5cf7de92fe34bd0a4701b11a4dba7ae384f2d0e33f185656107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6565dc4c35c618a239bdc2be6dd45a9057e83573c92c3fc816350eb014c822c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2667196228a62c452b7eaff0103e9bd92e88829c669c109b415b2fe28bb8cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.220883 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.220984 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.221004 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.221031 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.221049 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:29Z","lastTransitionTime":"2026-01-29T16:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.226067 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.243113 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8c765f3-89eb-4077-8829-03e86eb0c90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bec263dd306a333c63e0672bd5b5a5bf7cd7814c2c51bb480aac7c8e35591d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsqf5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ppngk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.275093 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b20f02-6c1e-4082-8233-8f06bda63195\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:11:16Z\\\",\\\"message\\\":\\\"] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:11:15.260241 6760 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 16:11:15.260645 6760 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0129 16:11:15.261304 6760 obj_retry.go:551] Creating *factory.egressNode crc took: 14.368203ms\\\\nI0129 16:11:15.261337 6760 factory.go:1336] Added *v1.Node event handler 7\\\\nI0129 16:11:15.261379 6760 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0129 16:11:15.261465 6760 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0129 16:11:15.261482 6760 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0129 16:11:15.261500 6760 factory.go:656] Stopping watch factory\\\\nI0129 16:11:15.261522 6760 handler.go:208] Removed *v1.Node event handler 7\\\\nI0129 16:11:15.261535 6760 handler.go:208] Removed *v1.Node event handler 2\\\\nI0129 16:11:15.261674 6760 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0129 16:11:15.261875 6760 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0129 16:11:15.261909 6760 ovnkube.go:599] Stopped ovnkube\\\\nI0129 16:11:15.261936 6760 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0129 16:11:15.262026 6760 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:11:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-sbnkt_openshift-ovn-kubernetes(04b20f02-6c1e-4082-8233-8f06bda63195)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vrsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-sbnkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.293402 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2932c3bd-04c7-4494-8d43-03c4524a353f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78109e9e498f3d640934a2c45faf27133d32ece9e35e42ed48ddba720fa7a10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc1d7b5f32fe8de4bfd5ba555838989293be538217424c86ea5cedadb8295f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvrtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tg8sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.314249 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.324211 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.324459 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.324627 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.324802 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.324981 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:29Z","lastTransitionTime":"2026-01-29T16:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.334662 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.366391 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b20fd8d-1ebb-47d0-8676-403b99dac1ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ec9086c2f128a0ee5f564692cf28086688a45cf9d6b328c33446ad8c6f2f4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19947ad8dfd71d43fffc2a5975a0d1663180736ca519dc2a1c4bafd17cbcc76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7164e1bbcac4ad832627cb9036637cc92f4b4285831e6a04abfd4fd0904e21a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://312a225d18b0eb9ae91f8a87303ef4c896c6aa7435fcf4485d8e2eda65a474ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37527bb32bfd42ed38400659a3b3df47f711df111fc318cd6ca42e4756408df7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93629f3fad3c74e2736307cf494b6301e63f371189d21fb879d22e1535a31a59\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ca06d57822f3e105b504240bdddd2d1b5c0b6650afcd0268cd1fa71766687d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:10:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vkc7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2cfxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.380918 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-c9jhc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f80aba4c-9372-4bea-b537-cbd9b0a3e972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc0e24a214494b56ffc8998f30296ecb7f846af6ba355b1a7ced0612f34143e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-th2m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-c9jhc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.396839 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2w92b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791456e8-8d95-4cdb-8fd1-d06a7586b328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwv7p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2w92b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.420727 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T16:10:08Z\\\",\\\"message\\\":\\\"W0129 16:09:57.450571 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 16:09:57.450796 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769702997 cert, and key in /tmp/serving-cert-976236263/serving-signer.crt, /tmp/serving-cert-976236263/serving-signer.key\\\\nI0129 16:09:57.732705 1 observer_polling.go:159] Starting file observer\\\\nW0129 16:09:57.737820 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 16:09:57.738179 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 16:09:57.739209 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-976236263/tls.crt::/tmp/serving-cert-976236263/tls.key\\\\\\\"\\\\nF0129 16:10:08.232982 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.429243 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.429326 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.429352 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.429384 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.429408 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:29Z","lastTransitionTime":"2026-01-29T16:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.452277 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56dc350a62212e8fe3f0e111610fd5fb645582503be465949ea0e97e9dab1a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44ea000eb0a41ef1ceab485441cd6ad2c665f81f0758597f166c693f23ae3315\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.469800 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-46dqc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f846b283-5468-4014-ba05-da5bfffa2ebc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4885f484b167e32a8d0767759d7c131508ad3c99a5019b237912064c8152510c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbd9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-46dqc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.491591 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b2ttm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89560008-8bdc-4640-af11-681d825e69d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c747ed61a18e27d63630395860ce896242426b1ce46ea5f9d00534b808804a58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T16:11:03Z\\\",\\\"message\\\":\\\"2026-01-29T16:10:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_27c11876-c5ae-4d35-b720-78ab09bfac0c\\\\n2026-01-29T16:10:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_27c11876-c5ae-4d35-b720-78ab09bfac0c to /host/opt/cni/bin/\\\\n2026-01-29T16:10:18Z [verbose] multus-daemon started\\\\n2026-01-29T16:10:18Z [verbose] Readiness Indicator file check\\\\n2026-01-29T16:11:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T16:10:16Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp6mh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:10:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b2ttm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.506084 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a65c2b0-9568-4a06-8073-93ec194b4ee0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89cb2f8d441042c4c95e3cc056f991565c18bd93dcb0d61f3735e2451ff439a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86530d623401c3cf5ebe44fcc1c2110a5f4dde059a5d27f74e118d5ca1df40dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86530d623401c3cf5ebe44fcc1c2110a5f4dde059a5d27f74e118d5ca1df40dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.532167 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.532224 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.532243 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.532281 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.532306 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:29Z","lastTransitionTime":"2026-01-29T16:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.536908 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"acb99202-b98d-4b54-bec6-9e6a90d5cbd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddd8695e3d9e8fbc9ce0cd1d83ee134a3cd1377940f4b8763ce0999b185f02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://988169eacf2bf87759260b2f9a1a8786b0bdfb3fd2c0b4f4ea2425d1eaa5ccd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7960e9e1ccf45e585a4a9610e1f5684caa0d939bc553335f1563ea5f1408346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c9bf119056f20eb63b1c55993b4e5e4fbce0ef1e1a0fc20c49047eb9c2af1b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9d42b2248c67787b28fa9139907de67ddd709e032ea2a495acbc0a5c2f2a109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b585d5aaa0cedba5a448db0c17eb71e468c5b8b091f0c61767217f5a949c8de7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f5ccea52820f488eae0c05e6a08b7e1ff0374f48484107886cf8e45e064965c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f73971e4ba98224ed462ff8491a7fd41dbb7c64a15149330a435d91eefd6d334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.558311 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61e7a70f-3b2d-4ee9-b9b9-160724395d19\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T16:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f890bfe59f701daca9b7eb40154d970341a9bdba499dc934e091e90a4a30c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee5dfdef978b46c3177ff0c01197e9d43fd086408ff2ed4f81199581bcf6a30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57c5cff6518bf6978dfdf9c3e8ec2fee7b23911d399d5467f91cda5da792f995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T16:09:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.580323 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13410896f482e837a870858f589327c25b1e85d9dd6f567853a8ff6aec87294b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.597024 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:10:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79d7d2d8c31924f92e593dd3fd079569b0dc714c7b9d517961ade71c470cf8b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T16:11:29Z is after 2025-08-24T17:21:41Z" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.635344 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.635412 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.635430 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.635455 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.635475 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:29Z","lastTransitionTime":"2026-01-29T16:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.738968 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.739052 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.739080 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.739113 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.739135 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:29Z","lastTransitionTime":"2026-01-29T16:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.842833 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.842896 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.842911 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.842960 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.842974 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:29Z","lastTransitionTime":"2026-01-29T16:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.864435 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 12:19:04.312450248 +0000 UTC Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.945824 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.945866 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.945876 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.945891 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:29 crc kubenswrapper[4714]: I0129 16:11:29.945904 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:29Z","lastTransitionTime":"2026-01-29T16:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.048903 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.049077 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.049097 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.049127 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.049148 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:30Z","lastTransitionTime":"2026-01-29T16:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.152147 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.152441 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.152518 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.152601 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.152657 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:30Z","lastTransitionTime":"2026-01-29T16:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.183385 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.183501 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.183538 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:30 crc kubenswrapper[4714]: E0129 16:11:30.183595 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:30 crc kubenswrapper[4714]: E0129 16:11:30.183895 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:30 crc kubenswrapper[4714]: E0129 16:11:30.184292 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.256165 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.256232 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.256250 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.256272 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.256290 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:30Z","lastTransitionTime":"2026-01-29T16:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.359334 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.359391 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.359409 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.359433 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.359454 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:30Z","lastTransitionTime":"2026-01-29T16:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.462778 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.462864 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.462889 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.462920 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.462988 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:30Z","lastTransitionTime":"2026-01-29T16:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.566207 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.566516 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.566601 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.566696 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.566783 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:30Z","lastTransitionTime":"2026-01-29T16:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.669895 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.670377 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.670561 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.670716 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.670962 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:30Z","lastTransitionTime":"2026-01-29T16:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.773739 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.773783 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.773795 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.773813 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.773827 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:30Z","lastTransitionTime":"2026-01-29T16:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.865016 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 06:56:06.535804265 +0000 UTC Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.876830 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.877147 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.877289 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.877429 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.877577 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:30Z","lastTransitionTime":"2026-01-29T16:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.980527 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.980976 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.981175 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.981393 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:30 crc kubenswrapper[4714]: I0129 16:11:30.981570 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:30Z","lastTransitionTime":"2026-01-29T16:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.085765 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.085852 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.086098 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.086148 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.086176 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:31Z","lastTransitionTime":"2026-01-29T16:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.184347 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:31 crc kubenswrapper[4714]: E0129 16:11:31.185623 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.189960 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.190068 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.190089 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.190153 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.190173 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:31Z","lastTransitionTime":"2026-01-29T16:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.293463 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.293534 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.293558 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.293584 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.293601 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:31Z","lastTransitionTime":"2026-01-29T16:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.396371 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.396442 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.396458 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.396481 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.396497 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:31Z","lastTransitionTime":"2026-01-29T16:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.500523 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.500614 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.500643 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.500685 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.500709 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:31Z","lastTransitionTime":"2026-01-29T16:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.604109 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.604184 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.604208 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.604241 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.604265 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:31Z","lastTransitionTime":"2026-01-29T16:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.706427 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.706473 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.706485 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.706500 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.706510 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:31Z","lastTransitionTime":"2026-01-29T16:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.809769 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.809862 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.809874 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.809893 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.809905 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:31Z","lastTransitionTime":"2026-01-29T16:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.865918 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 04:03:46.137696162 +0000 UTC Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.913224 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.913256 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.913264 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.913278 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:31 crc kubenswrapper[4714]: I0129 16:11:31.913286 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:31Z","lastTransitionTime":"2026-01-29T16:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.016656 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.016731 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.016756 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.016786 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.016803 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:32Z","lastTransitionTime":"2026-01-29T16:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.120020 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.120482 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.120688 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.120830 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.120997 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:32Z","lastTransitionTime":"2026-01-29T16:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.184057 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.184092 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:32 crc kubenswrapper[4714]: E0129 16:11:32.184264 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.184359 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:32 crc kubenswrapper[4714]: E0129 16:11:32.184406 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:32 crc kubenswrapper[4714]: E0129 16:11:32.184739 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.224800 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.224867 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.224886 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.224910 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.224966 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:32Z","lastTransitionTime":"2026-01-29T16:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.331563 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.331645 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.331671 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.331701 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.331720 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:32Z","lastTransitionTime":"2026-01-29T16:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.435812 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.435869 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.435880 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.435909 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.435923 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:32Z","lastTransitionTime":"2026-01-29T16:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.542066 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.542104 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.542117 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.542132 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.542144 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:32Z","lastTransitionTime":"2026-01-29T16:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.652686 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.652751 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.652763 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.652780 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.652791 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:32Z","lastTransitionTime":"2026-01-29T16:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.754485 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.754547 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.754557 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.754571 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.754582 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:32Z","lastTransitionTime":"2026-01-29T16:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.857059 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.857101 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.857110 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.857130 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.857142 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:32Z","lastTransitionTime":"2026-01-29T16:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.866174 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 07:38:31.408393991 +0000 UTC Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.880087 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.880131 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.880142 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.880160 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.880170 4714 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T16:11:32Z","lastTransitionTime":"2026-01-29T16:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.984417 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr"] Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.984919 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.986512 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.986596 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.988562 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 29 16:11:32 crc kubenswrapper[4714]: I0129 16:11:32.988707 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.008074 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-2cfxk" podStartSLOduration=79.008051164 podStartE2EDuration="1m19.008051164s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:11:33.007958141 +0000 UTC m=+99.528459271" watchObservedRunningTime="2026-01-29 16:11:33.008051164 +0000 UTC m=+99.528552284" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.033056 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-c9jhc" podStartSLOduration=79.033036883 podStartE2EDuration="1m19.033036883s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:11:33.018639388 +0000 UTC m=+99.539140508" watchObservedRunningTime="2026-01-29 16:11:33.033036883 +0000 UTC m=+99.553538003" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.070860 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-46dqc" podStartSLOduration=79.070842391 podStartE2EDuration="1m19.070842391s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:11:33.070750439 +0000 UTC m=+99.591251559" watchObservedRunningTime="2026-01-29 16:11:33.070842391 +0000 UTC m=+99.591343511" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.081234 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6122b3fa-a2a7-4328-a057-ee8692c5dc83-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dvwfr\" (UID: \"6122b3fa-a2a7-4328-a057-ee8692c5dc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.081270 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6122b3fa-a2a7-4328-a057-ee8692c5dc83-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dvwfr\" (UID: \"6122b3fa-a2a7-4328-a057-ee8692c5dc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.081317 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6122b3fa-a2a7-4328-a057-ee8692c5dc83-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dvwfr\" (UID: \"6122b3fa-a2a7-4328-a057-ee8692c5dc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.081336 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6122b3fa-a2a7-4328-a057-ee8692c5dc83-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dvwfr\" (UID: \"6122b3fa-a2a7-4328-a057-ee8692c5dc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.081359 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6122b3fa-a2a7-4328-a057-ee8692c5dc83-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dvwfr\" (UID: \"6122b3fa-a2a7-4328-a057-ee8692c5dc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.097620 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-b2ttm" podStartSLOduration=79.097604792 podStartE2EDuration="1m19.097604792s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:11:33.082752654 +0000 UTC m=+99.603253774" watchObservedRunningTime="2026-01-29 16:11:33.097604792 +0000 UTC m=+99.618105902" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.111256 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=80.111236234 podStartE2EDuration="1m20.111236234s" podCreationTimestamp="2026-01-29 16:10:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:11:33.098091296 +0000 UTC m=+99.618592416" watchObservedRunningTime="2026-01-29 16:11:33.111236234 +0000 UTC m=+99.631737364" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.140127 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=79.140110336 podStartE2EDuration="1m19.140110336s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:11:33.12463544 +0000 UTC m=+99.645136570" watchObservedRunningTime="2026-01-29 16:11:33.140110336 +0000 UTC m=+99.660611456" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.162020 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=26.162002646 podStartE2EDuration="26.162002646s" podCreationTimestamp="2026-01-29 16:11:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:11:33.160696408 +0000 UTC m=+99.681197528" watchObservedRunningTime="2026-01-29 16:11:33.162002646 +0000 UTC m=+99.682503766" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.182509 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6122b3fa-a2a7-4328-a057-ee8692c5dc83-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dvwfr\" (UID: \"6122b3fa-a2a7-4328-a057-ee8692c5dc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.182824 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6122b3fa-a2a7-4328-a057-ee8692c5dc83-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dvwfr\" (UID: \"6122b3fa-a2a7-4328-a057-ee8692c5dc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.182920 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6122b3fa-a2a7-4328-a057-ee8692c5dc83-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dvwfr\" (UID: \"6122b3fa-a2a7-4328-a057-ee8692c5dc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.183102 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6122b3fa-a2a7-4328-a057-ee8692c5dc83-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dvwfr\" (UID: \"6122b3fa-a2a7-4328-a057-ee8692c5dc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.183228 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6122b3fa-a2a7-4328-a057-ee8692c5dc83-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dvwfr\" (UID: \"6122b3fa-a2a7-4328-a057-ee8692c5dc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.183295 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.183447 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6122b3fa-a2a7-4328-a057-ee8692c5dc83-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dvwfr\" (UID: \"6122b3fa-a2a7-4328-a057-ee8692c5dc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.183580 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6122b3fa-a2a7-4328-a057-ee8692c5dc83-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dvwfr\" (UID: \"6122b3fa-a2a7-4328-a057-ee8692c5dc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.184044 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=77.1840359 podStartE2EDuration="1m17.1840359s" podCreationTimestamp="2026-01-29 16:10:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:11:33.183543006 +0000 UTC m=+99.704044136" watchObservedRunningTime="2026-01-29 16:11:33.1840359 +0000 UTC m=+99.704537020" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.184259 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6122b3fa-a2a7-4328-a057-ee8692c5dc83-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dvwfr\" (UID: \"6122b3fa-a2a7-4328-a057-ee8692c5dc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr" Jan 29 16:11:33 crc kubenswrapper[4714]: E0129 16:11:33.184375 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.189615 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6122b3fa-a2a7-4328-a057-ee8692c5dc83-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dvwfr\" (UID: \"6122b3fa-a2a7-4328-a057-ee8692c5dc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.199902 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6122b3fa-a2a7-4328-a057-ee8692c5dc83-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dvwfr\" (UID: \"6122b3fa-a2a7-4328-a057-ee8692c5dc83\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.220326 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podStartSLOduration=79.220309325 podStartE2EDuration="1m19.220309325s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:11:33.199268679 +0000 UTC m=+99.719769799" watchObservedRunningTime="2026-01-29 16:11:33.220309325 +0000 UTC m=+99.740810445" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.246999 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tg8sw" podStartSLOduration=78.246983642 podStartE2EDuration="1m18.246983642s" podCreationTimestamp="2026-01-29 16:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:11:33.24583672 +0000 UTC m=+99.766337840" watchObservedRunningTime="2026-01-29 16:11:33.246983642 +0000 UTC m=+99.767484762" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.267826 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=52.267805742 podStartE2EDuration="52.267805742s" podCreationTimestamp="2026-01-29 16:10:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:11:33.26739018 +0000 UTC m=+99.787891310" watchObservedRunningTime="2026-01-29 16:11:33.267805742 +0000 UTC m=+99.788306862" Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.301002 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr" Jan 29 16:11:33 crc kubenswrapper[4714]: W0129 16:11:33.317056 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6122b3fa_a2a7_4328_a057_ee8692c5dc83.slice/crio-ec683597ef95379a94031b5c628e9e0cf98cca1f2833bf03c7c91470ae7b3100 WatchSource:0}: Error finding container ec683597ef95379a94031b5c628e9e0cf98cca1f2833bf03c7c91470ae7b3100: Status 404 returned error can't find the container with id ec683597ef95379a94031b5c628e9e0cf98cca1f2833bf03c7c91470ae7b3100 Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.780236 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr" event={"ID":"6122b3fa-a2a7-4328-a057-ee8692c5dc83","Type":"ContainerStarted","Data":"ec683597ef95379a94031b5c628e9e0cf98cca1f2833bf03c7c91470ae7b3100"} Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.866377 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 11:55:54.017004966 +0000 UTC Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.866700 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Jan 29 16:11:33 crc kubenswrapper[4714]: I0129 16:11:33.875468 4714 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 29 16:11:34 crc kubenswrapper[4714]: I0129 16:11:34.093351 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs\") pod \"network-metrics-daemon-2w92b\" (UID: \"791456e8-8d95-4cdb-8fd1-d06a7586b328\") " pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:34 crc kubenswrapper[4714]: E0129 16:11:34.093500 4714 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 16:11:34 crc kubenswrapper[4714]: E0129 16:11:34.093548 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs podName:791456e8-8d95-4cdb-8fd1-d06a7586b328 nodeName:}" failed. No retries permitted until 2026-01-29 16:12:38.093532375 +0000 UTC m=+164.614033505 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs") pod "network-metrics-daemon-2w92b" (UID: "791456e8-8d95-4cdb-8fd1-d06a7586b328") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 16:11:34 crc kubenswrapper[4714]: I0129 16:11:34.183680 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:34 crc kubenswrapper[4714]: I0129 16:11:34.183736 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:34 crc kubenswrapper[4714]: E0129 16:11:34.184708 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:34 crc kubenswrapper[4714]: I0129 16:11:34.184806 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:34 crc kubenswrapper[4714]: E0129 16:11:34.184984 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:34 crc kubenswrapper[4714]: E0129 16:11:34.185257 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:34 crc kubenswrapper[4714]: I0129 16:11:34.787127 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr" event={"ID":"6122b3fa-a2a7-4328-a057-ee8692c5dc83","Type":"ContainerStarted","Data":"5d50732e6565a1ea71d386c131b1504da8f37e3b89c4683cff6fd99dd3fbb374"} Jan 29 16:11:34 crc kubenswrapper[4714]: I0129 16:11:34.804754 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dvwfr" podStartSLOduration=80.804734341 podStartE2EDuration="1m20.804734341s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:11:34.803004501 +0000 UTC m=+101.323505631" watchObservedRunningTime="2026-01-29 16:11:34.804734341 +0000 UTC m=+101.325235471" Jan 29 16:11:35 crc kubenswrapper[4714]: I0129 16:11:35.183684 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:35 crc kubenswrapper[4714]: E0129 16:11:35.183922 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:36 crc kubenswrapper[4714]: I0129 16:11:36.183564 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:36 crc kubenswrapper[4714]: I0129 16:11:36.183679 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:36 crc kubenswrapper[4714]: I0129 16:11:36.183588 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:36 crc kubenswrapper[4714]: E0129 16:11:36.183846 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:36 crc kubenswrapper[4714]: E0129 16:11:36.183982 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:36 crc kubenswrapper[4714]: E0129 16:11:36.184365 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:37 crc kubenswrapper[4714]: I0129 16:11:37.184077 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:37 crc kubenswrapper[4714]: E0129 16:11:37.184259 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:38 crc kubenswrapper[4714]: I0129 16:11:38.185263 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:38 crc kubenswrapper[4714]: I0129 16:11:38.185365 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:38 crc kubenswrapper[4714]: I0129 16:11:38.186218 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:38 crc kubenswrapper[4714]: E0129 16:11:38.186377 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:38 crc kubenswrapper[4714]: E0129 16:11:38.186653 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:38 crc kubenswrapper[4714]: E0129 16:11:38.186869 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:39 crc kubenswrapper[4714]: I0129 16:11:39.183427 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:39 crc kubenswrapper[4714]: E0129 16:11:39.184001 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:40 crc kubenswrapper[4714]: I0129 16:11:40.183151 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:40 crc kubenswrapper[4714]: I0129 16:11:40.183266 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:40 crc kubenswrapper[4714]: I0129 16:11:40.183583 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:40 crc kubenswrapper[4714]: E0129 16:11:40.183297 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:40 crc kubenswrapper[4714]: E0129 16:11:40.183478 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:40 crc kubenswrapper[4714]: E0129 16:11:40.183831 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:41 crc kubenswrapper[4714]: I0129 16:11:41.183762 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:41 crc kubenswrapper[4714]: E0129 16:11:41.184028 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:42 crc kubenswrapper[4714]: I0129 16:11:42.183815 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:42 crc kubenswrapper[4714]: I0129 16:11:42.183870 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:42 crc kubenswrapper[4714]: I0129 16:11:42.183842 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:42 crc kubenswrapper[4714]: E0129 16:11:42.184056 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:42 crc kubenswrapper[4714]: E0129 16:11:42.184263 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:42 crc kubenswrapper[4714]: E0129 16:11:42.184349 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:43 crc kubenswrapper[4714]: I0129 16:11:43.183831 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:43 crc kubenswrapper[4714]: E0129 16:11:43.184074 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:44 crc kubenswrapper[4714]: I0129 16:11:44.184073 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:44 crc kubenswrapper[4714]: I0129 16:11:44.184215 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:44 crc kubenswrapper[4714]: E0129 16:11:44.187297 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:44 crc kubenswrapper[4714]: I0129 16:11:44.187351 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:44 crc kubenswrapper[4714]: E0129 16:11:44.187535 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:44 crc kubenswrapper[4714]: E0129 16:11:44.187707 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:44 crc kubenswrapper[4714]: I0129 16:11:44.189016 4714 scope.go:117] "RemoveContainer" containerID="95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365" Jan 29 16:11:44 crc kubenswrapper[4714]: E0129 16:11:44.189324 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-sbnkt_openshift-ovn-kubernetes(04b20f02-6c1e-4082-8233-8f06bda63195)\"" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" Jan 29 16:11:45 crc kubenswrapper[4714]: I0129 16:11:45.184012 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:45 crc kubenswrapper[4714]: E0129 16:11:45.184254 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:46 crc kubenswrapper[4714]: I0129 16:11:46.183591 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:46 crc kubenswrapper[4714]: I0129 16:11:46.183630 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:46 crc kubenswrapper[4714]: E0129 16:11:46.183895 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:46 crc kubenswrapper[4714]: I0129 16:11:46.183918 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:46 crc kubenswrapper[4714]: E0129 16:11:46.184034 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:46 crc kubenswrapper[4714]: E0129 16:11:46.184150 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:47 crc kubenswrapper[4714]: I0129 16:11:47.183924 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:47 crc kubenswrapper[4714]: E0129 16:11:47.184228 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:48 crc kubenswrapper[4714]: I0129 16:11:48.183781 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:48 crc kubenswrapper[4714]: I0129 16:11:48.183899 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:48 crc kubenswrapper[4714]: I0129 16:11:48.183998 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:48 crc kubenswrapper[4714]: E0129 16:11:48.184023 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:48 crc kubenswrapper[4714]: E0129 16:11:48.184226 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:48 crc kubenswrapper[4714]: E0129 16:11:48.184268 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:49 crc kubenswrapper[4714]: I0129 16:11:49.184074 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:49 crc kubenswrapper[4714]: E0129 16:11:49.184239 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:50 crc kubenswrapper[4714]: I0129 16:11:50.190045 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:50 crc kubenswrapper[4714]: E0129 16:11:50.190178 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:50 crc kubenswrapper[4714]: I0129 16:11:50.190049 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:50 crc kubenswrapper[4714]: E0129 16:11:50.190394 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:50 crc kubenswrapper[4714]: I0129 16:11:50.190466 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:50 crc kubenswrapper[4714]: E0129 16:11:50.190526 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:50 crc kubenswrapper[4714]: I0129 16:11:50.838179 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b2ttm_89560008-8bdc-4640-af11-681d825e69d4/kube-multus/1.log" Jan 29 16:11:50 crc kubenswrapper[4714]: I0129 16:11:50.838771 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b2ttm_89560008-8bdc-4640-af11-681d825e69d4/kube-multus/0.log" Jan 29 16:11:50 crc kubenswrapper[4714]: I0129 16:11:50.838829 4714 generic.go:334] "Generic (PLEG): container finished" podID="89560008-8bdc-4640-af11-681d825e69d4" containerID="c747ed61a18e27d63630395860ce896242426b1ce46ea5f9d00534b808804a58" exitCode=1 Jan 29 16:11:50 crc kubenswrapper[4714]: I0129 16:11:50.838870 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b2ttm" event={"ID":"89560008-8bdc-4640-af11-681d825e69d4","Type":"ContainerDied","Data":"c747ed61a18e27d63630395860ce896242426b1ce46ea5f9d00534b808804a58"} Jan 29 16:11:50 crc kubenswrapper[4714]: I0129 16:11:50.838909 4714 scope.go:117] "RemoveContainer" containerID="a63101a231660b105a82b67269b53217cac5e28d81c8a9e123d259779a76b84a" Jan 29 16:11:50 crc kubenswrapper[4714]: I0129 16:11:50.839458 4714 scope.go:117] "RemoveContainer" containerID="c747ed61a18e27d63630395860ce896242426b1ce46ea5f9d00534b808804a58" Jan 29 16:11:50 crc kubenswrapper[4714]: E0129 16:11:50.839645 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-b2ttm_openshift-multus(89560008-8bdc-4640-af11-681d825e69d4)\"" pod="openshift-multus/multus-b2ttm" podUID="89560008-8bdc-4640-af11-681d825e69d4" Jan 29 16:11:51 crc kubenswrapper[4714]: I0129 16:11:51.183387 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:51 crc kubenswrapper[4714]: E0129 16:11:51.183736 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:51 crc kubenswrapper[4714]: I0129 16:11:51.844020 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b2ttm_89560008-8bdc-4640-af11-681d825e69d4/kube-multus/1.log" Jan 29 16:11:52 crc kubenswrapper[4714]: I0129 16:11:52.184082 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:52 crc kubenswrapper[4714]: I0129 16:11:52.184144 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:52 crc kubenswrapper[4714]: I0129 16:11:52.184112 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:52 crc kubenswrapper[4714]: E0129 16:11:52.184243 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:52 crc kubenswrapper[4714]: E0129 16:11:52.184430 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:52 crc kubenswrapper[4714]: E0129 16:11:52.184520 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:53 crc kubenswrapper[4714]: I0129 16:11:53.183150 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:53 crc kubenswrapper[4714]: E0129 16:11:53.183292 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:54 crc kubenswrapper[4714]: I0129 16:11:54.184240 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:54 crc kubenswrapper[4714]: I0129 16:11:54.184239 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:54 crc kubenswrapper[4714]: I0129 16:11:54.184351 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:54 crc kubenswrapper[4714]: E0129 16:11:54.192438 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:54 crc kubenswrapper[4714]: E0129 16:11:54.193170 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:54 crc kubenswrapper[4714]: E0129 16:11:54.194040 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:54 crc kubenswrapper[4714]: E0129 16:11:54.207085 4714 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Jan 29 16:11:54 crc kubenswrapper[4714]: E0129 16:11:54.280364 4714 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 29 16:11:55 crc kubenswrapper[4714]: I0129 16:11:55.183246 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:55 crc kubenswrapper[4714]: E0129 16:11:55.183505 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:56 crc kubenswrapper[4714]: I0129 16:11:56.183388 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:56 crc kubenswrapper[4714]: I0129 16:11:56.183453 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:56 crc kubenswrapper[4714]: E0129 16:11:56.183560 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:56 crc kubenswrapper[4714]: I0129 16:11:56.183598 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:56 crc kubenswrapper[4714]: E0129 16:11:56.183697 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:56 crc kubenswrapper[4714]: E0129 16:11:56.183763 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:57 crc kubenswrapper[4714]: I0129 16:11:57.183665 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:57 crc kubenswrapper[4714]: E0129 16:11:57.184012 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:57 crc kubenswrapper[4714]: I0129 16:11:57.185153 4714 scope.go:117] "RemoveContainer" containerID="95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365" Jan 29 16:11:57 crc kubenswrapper[4714]: I0129 16:11:57.868317 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-sbnkt_04b20f02-6c1e-4082-8233-8f06bda63195/ovnkube-controller/3.log" Jan 29 16:11:57 crc kubenswrapper[4714]: I0129 16:11:57.871662 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerStarted","Data":"00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6"} Jan 29 16:11:57 crc kubenswrapper[4714]: I0129 16:11:57.872153 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:11:57 crc kubenswrapper[4714]: I0129 16:11:57.925120 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" podStartSLOduration=103.925093306 podStartE2EDuration="1m43.925093306s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:11:57.922261564 +0000 UTC m=+124.442762724" watchObservedRunningTime="2026-01-29 16:11:57.925093306 +0000 UTC m=+124.445594466" Jan 29 16:11:58 crc kubenswrapper[4714]: I0129 16:11:58.183656 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:58 crc kubenswrapper[4714]: I0129 16:11:58.183710 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:11:58 crc kubenswrapper[4714]: I0129 16:11:58.183801 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:11:58 crc kubenswrapper[4714]: E0129 16:11:58.184026 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:58 crc kubenswrapper[4714]: E0129 16:11:58.184165 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:11:58 crc kubenswrapper[4714]: E0129 16:11:58.184269 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:11:58 crc kubenswrapper[4714]: I0129 16:11:58.207412 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-2w92b"] Jan 29 16:11:58 crc kubenswrapper[4714]: I0129 16:11:58.876452 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:11:58 crc kubenswrapper[4714]: E0129 16:11:58.877138 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:11:59 crc kubenswrapper[4714]: I0129 16:11:59.183247 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:11:59 crc kubenswrapper[4714]: E0129 16:11:59.183420 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:11:59 crc kubenswrapper[4714]: E0129 16:11:59.281795 4714 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 29 16:12:00 crc kubenswrapper[4714]: I0129 16:12:00.184131 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:12:00 crc kubenswrapper[4714]: I0129 16:12:00.184251 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:12:00 crc kubenswrapper[4714]: E0129 16:12:00.184309 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:12:00 crc kubenswrapper[4714]: E0129 16:12:00.184471 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:12:00 crc kubenswrapper[4714]: I0129 16:12:00.184736 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:12:00 crc kubenswrapper[4714]: E0129 16:12:00.184848 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:12:01 crc kubenswrapper[4714]: I0129 16:12:01.184076 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:12:01 crc kubenswrapper[4714]: E0129 16:12:01.184325 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:12:02 crc kubenswrapper[4714]: I0129 16:12:02.183576 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:12:02 crc kubenswrapper[4714]: I0129 16:12:02.183634 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:12:02 crc kubenswrapper[4714]: E0129 16:12:02.183842 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:12:02 crc kubenswrapper[4714]: I0129 16:12:02.183909 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:12:02 crc kubenswrapper[4714]: E0129 16:12:02.184166 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:12:02 crc kubenswrapper[4714]: E0129 16:12:02.184352 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:12:02 crc kubenswrapper[4714]: I0129 16:12:02.277352 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:12:03 crc kubenswrapper[4714]: I0129 16:12:03.183309 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:12:03 crc kubenswrapper[4714]: E0129 16:12:03.183494 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:12:04 crc kubenswrapper[4714]: I0129 16:12:04.184302 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:12:04 crc kubenswrapper[4714]: I0129 16:12:04.184431 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:12:04 crc kubenswrapper[4714]: E0129 16:12:04.184571 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:12:04 crc kubenswrapper[4714]: I0129 16:12:04.184623 4714 scope.go:117] "RemoveContainer" containerID="c747ed61a18e27d63630395860ce896242426b1ce46ea5f9d00534b808804a58" Jan 29 16:12:04 crc kubenswrapper[4714]: E0129 16:12:04.184678 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:12:04 crc kubenswrapper[4714]: I0129 16:12:04.185299 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:12:04 crc kubenswrapper[4714]: E0129 16:12:04.185466 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:12:04 crc kubenswrapper[4714]: E0129 16:12:04.283147 4714 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 29 16:12:04 crc kubenswrapper[4714]: I0129 16:12:04.901823 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b2ttm_89560008-8bdc-4640-af11-681d825e69d4/kube-multus/1.log" Jan 29 16:12:04 crc kubenswrapper[4714]: I0129 16:12:04.901913 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b2ttm" event={"ID":"89560008-8bdc-4640-af11-681d825e69d4","Type":"ContainerStarted","Data":"e21aab3b653d9b1f38d58e9c32cbfb8988660ecb96eec4099a6536e09747d8fb"} Jan 29 16:12:05 crc kubenswrapper[4714]: I0129 16:12:05.183614 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:12:05 crc kubenswrapper[4714]: E0129 16:12:05.183780 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:12:06 crc kubenswrapper[4714]: I0129 16:12:06.183646 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:12:06 crc kubenswrapper[4714]: E0129 16:12:06.183775 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:12:06 crc kubenswrapper[4714]: I0129 16:12:06.183839 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:12:06 crc kubenswrapper[4714]: I0129 16:12:06.183670 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:12:06 crc kubenswrapper[4714]: E0129 16:12:06.184057 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:12:06 crc kubenswrapper[4714]: E0129 16:12:06.183963 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:12:07 crc kubenswrapper[4714]: I0129 16:12:07.183250 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:12:07 crc kubenswrapper[4714]: E0129 16:12:07.183500 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:12:08 crc kubenswrapper[4714]: I0129 16:12:08.183659 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:12:08 crc kubenswrapper[4714]: I0129 16:12:08.183714 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:12:08 crc kubenswrapper[4714]: I0129 16:12:08.185310 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:12:08 crc kubenswrapper[4714]: E0129 16:12:08.185653 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 16:12:08 crc kubenswrapper[4714]: E0129 16:12:08.185756 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2w92b" podUID="791456e8-8d95-4cdb-8fd1-d06a7586b328" Jan 29 16:12:08 crc kubenswrapper[4714]: E0129 16:12:08.186049 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 16:12:09 crc kubenswrapper[4714]: I0129 16:12:09.184234 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:12:09 crc kubenswrapper[4714]: E0129 16:12:09.184405 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 16:12:10 crc kubenswrapper[4714]: I0129 16:12:10.183430 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:12:10 crc kubenswrapper[4714]: I0129 16:12:10.183491 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:12:10 crc kubenswrapper[4714]: I0129 16:12:10.184140 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:12:10 crc kubenswrapper[4714]: I0129 16:12:10.186294 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 29 16:12:10 crc kubenswrapper[4714]: I0129 16:12:10.188990 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 29 16:12:10 crc kubenswrapper[4714]: I0129 16:12:10.189220 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 29 16:12:10 crc kubenswrapper[4714]: I0129 16:12:10.189773 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 29 16:12:11 crc kubenswrapper[4714]: I0129 16:12:11.183808 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:12:11 crc kubenswrapper[4714]: I0129 16:12:11.185630 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 29 16:12:11 crc kubenswrapper[4714]: I0129 16:12:11.187033 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.630288 4714 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.685488 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-xvrxj"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.686328 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xvrxj" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.687549 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jb6jw"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.688497 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-jb6jw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.689565 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.689700 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.689733 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.690083 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.690148 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.693893 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.693911 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.694079 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.694075 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-kvp9d"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.695250 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-kvp9d" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.696785 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xlczd"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.697818 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.699272 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vcj84"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.699988 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vcj84" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.702571 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.703760 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.704193 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6jl75"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.711684 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.712024 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.726284 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4dn69"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.726703 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-m2g9h"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.727055 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-z4h55"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.727595 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.727704 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.727664 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-z4h55" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.728194 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4dn69" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.728564 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.728587 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.728782 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.728920 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.728960 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.729082 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.729301 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.729604 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.729722 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.729800 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.729822 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.729843 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.730024 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.730038 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.729723 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.730183 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.730408 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.730478 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.730416 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.730542 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.730488 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.730700 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-99knh"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.730614 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.730648 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.731628 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-99knh" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.732116 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.732224 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.732410 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.732453 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.732489 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.732576 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.732617 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-fn75b"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.733162 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-fn75b" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.732624 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.733611 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.732650 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.732660 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.732687 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.738980 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-h8b4r"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.739684 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-nh2m9"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.739987 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp5md"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.740146 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.740361 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.740468 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp5md" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.741081 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.741089 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.741674 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.741869 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.742061 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.742169 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.742258 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.742347 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.742439 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.742518 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.742603 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.742684 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.742796 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.742859 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.743254 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.748765 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.749059 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gnjmm"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.751752 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.760505 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.760782 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.761159 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-lz6mw"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.762114 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-nf7jb"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.763244 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.763373 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.764416 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.764954 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.765209 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.766038 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.766376 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.767316 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.767738 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.768263 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-sv7xw"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.768434 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nf7jb" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.769108 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.769356 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mrprd"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.770096 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.768280 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.768877 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.769269 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.769563 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.774260 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.774531 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.774636 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.774820 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.775134 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.775307 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.775966 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.795291 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.795533 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.795613 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.795885 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.796081 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.796251 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mrprd" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.796367 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.796477 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-sv7xw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.797111 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.797453 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2lpzx"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.797761 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.797921 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2lpzx" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.797963 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.798388 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.799095 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.799172 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.799258 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.801344 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.801637 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.801773 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.801917 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802077 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802423 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802464 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802496 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnhkm\" (UniqueName: \"kubernetes.io/projected/c779f8ba-7614-49f1-be6d-a9e316ec59ba-kube-api-access-rnhkm\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802522 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5b5da98c-0704-41c7-8563-707f7af93f41-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-99knh\" (UID: \"5b5da98c-0704-41c7-8563-707f7af93f41\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-99knh" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802546 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f3e2f962-69e3-4008-a45f-5c35677f7f36-machine-approver-tls\") pod \"machine-approver-56656f9798-xvrxj\" (UID: \"f3e2f962-69e3-4008-a45f-5c35677f7f36\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xvrxj" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802571 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92-config\") pod \"machine-api-operator-5694c8668f-z4h55\" (UID: \"bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z4h55" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802596 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99bab267-639b-48b1-abc4-8c0373200a39-config\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802618 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802641 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92-images\") pod \"machine-api-operator-5694c8668f-z4h55\" (UID: \"bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z4h55" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802666 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec-serving-cert\") pod \"console-operator-58897d9998-kvp9d\" (UID: \"ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec\") " pod="openshift-console-operator/console-operator-58897d9998-kvp9d" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802688 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0387af3d-8796-46b0-9282-9ecbda7fe3a7-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp5md\" (UID: \"0387af3d-8796-46b0-9282-9ecbda7fe3a7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp5md" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802721 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b25d77ec-57de-4c2a-b534-e98bf149b92a-serving-cert\") pod \"authentication-operator-69f744f599-jb6jw\" (UID: \"b25d77ec-57de-4c2a-b534-e98bf149b92a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jb6jw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802743 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/99bab267-639b-48b1-abc4-8c0373200a39-audit-dir\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802750 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802768 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec-config\") pod \"console-operator-58897d9998-kvp9d\" (UID: \"ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec\") " pod="openshift-console-operator/console-operator-58897d9998-kvp9d" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802789 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hv4x\" (UniqueName: \"kubernetes.io/projected/f3e2f962-69e3-4008-a45f-5c35677f7f36-kube-api-access-5hv4x\") pod \"machine-approver-56656f9798-xvrxj\" (UID: \"f3e2f962-69e3-4008-a45f-5c35677f7f36\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xvrxj" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802812 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97cfbecd-36ef-409b-94e9-f607a1fa2c42-config\") pod \"etcd-operator-b45778765-nh2m9\" (UID: \"97cfbecd-36ef-409b-94e9-f607a1fa2c42\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802834 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/6de35940-bef4-4dfa-9a83-08ba29d73399-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-4dn69\" (UID: \"6de35940-bef4-4dfa-9a83-08ba29d73399\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4dn69" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802872 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2rbv\" (UniqueName: \"kubernetes.io/projected/99bab267-639b-48b1-abc4-8c0373200a39-kube-api-access-f2rbv\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802897 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0e2a789d-6a90-4d60-881e-9562cd92e0a7-console-config\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.802925 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f3e2f962-69e3-4008-a45f-5c35677f7f36-auth-proxy-config\") pod \"machine-approver-56656f9798-xvrxj\" (UID: \"f3e2f962-69e3-4008-a45f-5c35677f7f36\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xvrxj" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803156 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b25d77ec-57de-4c2a-b534-e98bf149b92a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jb6jw\" (UID: \"b25d77ec-57de-4c2a-b534-e98bf149b92a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jb6jw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803196 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-z4h55\" (UID: \"bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z4h55" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803241 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/832097a5-4691-42b6-99cc-38679071d5ee-audit-dir\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803265 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803288 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knzbq\" (UniqueName: \"kubernetes.io/projected/8f71ba3e-c687-4ff7-9475-1e18ded764f6-kube-api-access-knzbq\") pod \"openshift-config-operator-7777fb866f-dwsm5\" (UID: \"8f71ba3e-c687-4ff7-9475-1e18ded764f6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803310 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c779f8ba-7614-49f1-be6d-a9e316ec59ba-serving-cert\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803334 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c2d0611-58f8-4a7e-8280-361c80d62802-config\") pod \"controller-manager-879f6c89f-xlczd\" (UID: \"3c2d0611-58f8-4a7e-8280-361c80d62802\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803361 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srhd7\" (UniqueName: \"kubernetes.io/projected/fbfdd647-1d64-4d35-9af2-6dee52b4c860-kube-api-access-srhd7\") pod \"route-controller-manager-6576b87f9c-m2qxw\" (UID: \"fbfdd647-1d64-4d35-9af2-6dee52b4c860\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803386 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99cl5\" (UniqueName: \"kubernetes.io/projected/b25d77ec-57de-4c2a-b534-e98bf149b92a-kube-api-access-99cl5\") pod \"authentication-operator-69f744f599-jb6jw\" (UID: \"b25d77ec-57de-4c2a-b534-e98bf149b92a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jb6jw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803411 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c779f8ba-7614-49f1-be6d-a9e316ec59ba-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803446 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/99bab267-639b-48b1-abc4-8c0373200a39-etcd-client\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803467 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-audit-policies\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803488 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803510 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3e2f962-69e3-4008-a45f-5c35677f7f36-config\") pod \"machine-approver-56656f9798-xvrxj\" (UID: \"f3e2f962-69e3-4008-a45f-5c35677f7f36\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xvrxj" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803532 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b25d77ec-57de-4c2a-b534-e98bf149b92a-service-ca-bundle\") pod \"authentication-operator-69f744f599-jb6jw\" (UID: \"b25d77ec-57de-4c2a-b534-e98bf149b92a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jb6jw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803554 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb7zc\" (UniqueName: \"kubernetes.io/projected/eacb9f84-018a-4f64-b211-c9bedce50b9e-kube-api-access-sb7zc\") pod \"openshift-apiserver-operator-796bbdcf4f-vcj84\" (UID: \"eacb9f84-018a-4f64-b211-c9bedce50b9e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vcj84" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803573 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/99bab267-639b-48b1-abc4-8c0373200a39-encryption-config\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803595 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803621 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbfdd647-1d64-4d35-9af2-6dee52b4c860-config\") pod \"route-controller-manager-6576b87f9c-m2qxw\" (UID: \"fbfdd647-1d64-4d35-9af2-6dee52b4c860\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803641 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b25d77ec-57de-4c2a-b534-e98bf149b92a-config\") pod \"authentication-operator-69f744f599-jb6jw\" (UID: \"b25d77ec-57de-4c2a-b534-e98bf149b92a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jb6jw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803662 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3c2d0611-58f8-4a7e-8280-361c80d62802-client-ca\") pod \"controller-manager-879f6c89f-xlczd\" (UID: \"3c2d0611-58f8-4a7e-8280-361c80d62802\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803688 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6de35940-bef4-4dfa-9a83-08ba29d73399-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-4dn69\" (UID: \"6de35940-bef4-4dfa-9a83-08ba29d73399\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4dn69" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803711 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/99bab267-639b-48b1-abc4-8c0373200a39-image-import-ca\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803733 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8nmp\" (UniqueName: \"kubernetes.io/projected/832097a5-4691-42b6-99cc-38679071d5ee-kube-api-access-l8nmp\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803758 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/99bab267-639b-48b1-abc4-8c0373200a39-node-pullsecrets\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803779 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803802 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eacb9f84-018a-4f64-b211-c9bedce50b9e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-vcj84\" (UID: \"eacb9f84-018a-4f64-b211-c9bedce50b9e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vcj84" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803822 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99bab267-639b-48b1-abc4-8c0373200a39-serving-cert\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803841 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fbfdd647-1d64-4d35-9af2-6dee52b4c860-client-ca\") pod \"route-controller-manager-6576b87f9c-m2qxw\" (UID: \"fbfdd647-1d64-4d35-9af2-6dee52b4c860\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803859 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0e2a789d-6a90-4d60-881e-9562cd92e0a7-service-ca\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803879 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0e2a789d-6a90-4d60-881e-9562cd92e0a7-console-serving-cert\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803900 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8tm6\" (UniqueName: \"kubernetes.io/projected/bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92-kube-api-access-k8tm6\") pod \"machine-api-operator-5694c8668f-z4h55\" (UID: \"bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z4h55" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.803919 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6de35940-bef4-4dfa-9a83-08ba29d73399-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-4dn69\" (UID: \"6de35940-bef4-4dfa-9a83-08ba29d73399\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4dn69" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.815387 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.816526 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8f71ba3e-c687-4ff7-9475-1e18ded764f6-available-featuregates\") pod \"openshift-config-operator-7777fb866f-dwsm5\" (UID: \"8f71ba3e-c687-4ff7-9475-1e18ded764f6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.816565 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmd5b\" (UniqueName: \"kubernetes.io/projected/5b5da98c-0704-41c7-8563-707f7af93f41-kube-api-access-hmd5b\") pod \"cluster-samples-operator-665b6dd947-99knh\" (UID: \"5b5da98c-0704-41c7-8563-707f7af93f41\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-99knh" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.816595 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec-trusted-ca\") pod \"console-operator-58897d9998-kvp9d\" (UID: \"ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec\") " pod="openshift-console-operator/console-operator-58897d9998-kvp9d" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.816624 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c779f8ba-7614-49f1-be6d-a9e316ec59ba-audit-policies\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.816649 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.816678 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0e2a789d-6a90-4d60-881e-9562cd92e0a7-oauth-serving-cert\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.816700 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f71ba3e-c687-4ff7-9475-1e18ded764f6-serving-cert\") pod \"openshift-config-operator-7777fb866f-dwsm5\" (UID: \"8f71ba3e-c687-4ff7-9475-1e18ded764f6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.816721 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c779f8ba-7614-49f1-be6d-a9e316ec59ba-etcd-client\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.816748 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x5xk\" (UniqueName: \"kubernetes.io/projected/ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec-kube-api-access-4x5xk\") pod \"console-operator-58897d9998-kvp9d\" (UID: \"ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec\") " pod="openshift-console-operator/console-operator-58897d9998-kvp9d" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.816772 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/97cfbecd-36ef-409b-94e9-f607a1fa2c42-etcd-service-ca\") pod \"etcd-operator-b45778765-nh2m9\" (UID: \"97cfbecd-36ef-409b-94e9-f607a1fa2c42\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.816793 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0387af3d-8796-46b0-9282-9ecbda7fe3a7-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp5md\" (UID: \"0387af3d-8796-46b0-9282-9ecbda7fe3a7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp5md" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.816818 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/97cfbecd-36ef-409b-94e9-f607a1fa2c42-etcd-ca\") pod \"etcd-operator-b45778765-nh2m9\" (UID: \"97cfbecd-36ef-409b-94e9-f607a1fa2c42\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.816842 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c779f8ba-7614-49f1-be6d-a9e316ec59ba-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.816865 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/99bab267-639b-48b1-abc4-8c0373200a39-etcd-serving-ca\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.816886 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk8rh\" (UniqueName: \"kubernetes.io/projected/3c2d0611-58f8-4a7e-8280-361c80d62802-kube-api-access-pk8rh\") pod \"controller-manager-879f6c89f-xlczd\" (UID: \"3c2d0611-58f8-4a7e-8280-361c80d62802\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.816908 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c779f8ba-7614-49f1-be6d-a9e316ec59ba-audit-dir\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.816949 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0e2a789d-6a90-4d60-881e-9562cd92e0a7-console-oauth-config\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.816971 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c779f8ba-7614-49f1-be6d-a9e316ec59ba-encryption-config\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.816993 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8qsw\" (UniqueName: \"kubernetes.io/projected/6de35940-bef4-4dfa-9a83-08ba29d73399-kube-api-access-j8qsw\") pod \"cluster-image-registry-operator-dc59b4c8b-4dn69\" (UID: \"6de35940-bef4-4dfa-9a83-08ba29d73399\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4dn69" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.817014 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.817037 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97cfbecd-36ef-409b-94e9-f607a1fa2c42-serving-cert\") pod \"etcd-operator-b45778765-nh2m9\" (UID: \"97cfbecd-36ef-409b-94e9-f607a1fa2c42\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.817063 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fbfdd647-1d64-4d35-9af2-6dee52b4c860-serving-cert\") pod \"route-controller-manager-6576b87f9c-m2qxw\" (UID: \"fbfdd647-1d64-4d35-9af2-6dee52b4c860\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.817085 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e2a789d-6a90-4d60-881e-9562cd92e0a7-trusted-ca-bundle\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.817110 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fbcn\" (UniqueName: \"kubernetes.io/projected/42b66dc3-a385-4350-a943-50f062da35f7-kube-api-access-2fbcn\") pod \"downloads-7954f5f757-fn75b\" (UID: \"42b66dc3-a385-4350-a943-50f062da35f7\") " pod="openshift-console/downloads-7954f5f757-fn75b" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.817132 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99bab267-639b-48b1-abc4-8c0373200a39-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.817153 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.817175 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.817212 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/97cfbecd-36ef-409b-94e9-f607a1fa2c42-etcd-client\") pod \"etcd-operator-b45778765-nh2m9\" (UID: \"97cfbecd-36ef-409b-94e9-f607a1fa2c42\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.817233 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3c2d0611-58f8-4a7e-8280-361c80d62802-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xlczd\" (UID: \"3c2d0611-58f8-4a7e-8280-361c80d62802\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.817256 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0387af3d-8796-46b0-9282-9ecbda7fe3a7-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp5md\" (UID: \"0387af3d-8796-46b0-9282-9ecbda7fe3a7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp5md" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.817281 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bn272\" (UniqueName: \"kubernetes.io/projected/0e2a789d-6a90-4d60-881e-9562cd92e0a7-kube-api-access-bn272\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.817306 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c2d0611-58f8-4a7e-8280-361c80d62802-serving-cert\") pod \"controller-manager-879f6c89f-xlczd\" (UID: \"3c2d0611-58f8-4a7e-8280-361c80d62802\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.817328 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eacb9f84-018a-4f64-b211-c9bedce50b9e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-vcj84\" (UID: \"eacb9f84-018a-4f64-b211-c9bedce50b9e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vcj84" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.817351 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjjl8\" (UniqueName: \"kubernetes.io/projected/97cfbecd-36ef-409b-94e9-f607a1fa2c42-kube-api-access-rjjl8\") pod \"etcd-operator-b45778765-nh2m9\" (UID: \"97cfbecd-36ef-409b-94e9-f607a1fa2c42\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.817372 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/99bab267-639b-48b1-abc4-8c0373200a39-audit\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.817262 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.827031 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.827574 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.828378 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.828595 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.828836 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.842814 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-l2t56"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.843613 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.844062 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ch6wr"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.844087 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.844553 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ch6wr" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.844690 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xtzbx"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.845314 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xtzbx" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.846199 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jzg2m"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.846749 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jzg2m" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.850194 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.850508 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.851838 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chzp2"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.852421 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chzp2" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.853339 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.868392 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-44gfk"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.869641 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-zkbcz"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.870813 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sq9mx"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.871704 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sq9mx" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.872055 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-44gfk" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.872303 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-zkbcz" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.874595 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.874868 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.876111 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.878987 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.879893 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pbrmk"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.880680 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.881144 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pbrmk" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.881182 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ljnh7"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.882429 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ljnh7" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.882696 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-jcdhl"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.885130 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-jcdhl" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.891022 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4dn69"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.891080 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jb6jw"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.892757 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.893099 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-kfqcf"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.903751 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xlczd"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.903871 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-kfqcf" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.906969 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.909164 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-m2g9h"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.909978 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.912035 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-fn75b"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.915838 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ch6wr"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.919010 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x5xk\" (UniqueName: \"kubernetes.io/projected/ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec-kube-api-access-4x5xk\") pod \"console-operator-58897d9998-kvp9d\" (UID: \"ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec\") " pod="openshift-console-operator/console-operator-58897d9998-kvp9d" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.919121 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/97cfbecd-36ef-409b-94e9-f607a1fa2c42-etcd-service-ca\") pod \"etcd-operator-b45778765-nh2m9\" (UID: \"97cfbecd-36ef-409b-94e9-f607a1fa2c42\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.919275 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0387af3d-8796-46b0-9282-9ecbda7fe3a7-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp5md\" (UID: \"0387af3d-8796-46b0-9282-9ecbda7fe3a7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp5md" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.919367 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45cdw\" (UniqueName: \"kubernetes.io/projected/2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96-kube-api-access-45cdw\") pod \"machine-config-controller-84d6567774-nf7jb\" (UID: \"2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nf7jb" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.919483 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/97cfbecd-36ef-409b-94e9-f607a1fa2c42-etcd-ca\") pod \"etcd-operator-b45778765-nh2m9\" (UID: \"97cfbecd-36ef-409b-94e9-f607a1fa2c42\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.919566 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c779f8ba-7614-49f1-be6d-a9e316ec59ba-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.919636 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/99bab267-639b-48b1-abc4-8c0373200a39-etcd-serving-ca\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.919716 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk8rh\" (UniqueName: \"kubernetes.io/projected/3c2d0611-58f8-4a7e-8280-361c80d62802-kube-api-access-pk8rh\") pod \"controller-manager-879f6c89f-xlczd\" (UID: \"3c2d0611-58f8-4a7e-8280-361c80d62802\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.919785 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0e2a789d-6a90-4d60-881e-9562cd92e0a7-console-oauth-config\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.919848 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c779f8ba-7614-49f1-be6d-a9e316ec59ba-encryption-config\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.919914 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c779f8ba-7614-49f1-be6d-a9e316ec59ba-audit-dir\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.920004 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-nf7jb\" (UID: \"2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nf7jb" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.920147 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwvc4\" (UniqueName: \"kubernetes.io/projected/554abf87-b1ba-45b1-8130-95b40da3b8bf-kube-api-access-zwvc4\") pod \"openshift-controller-manager-operator-756b6f6bc6-mrprd\" (UID: \"554abf87-b1ba-45b1-8130-95b40da3b8bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mrprd" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.920221 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97cfbecd-36ef-409b-94e9-f607a1fa2c42-serving-cert\") pod \"etcd-operator-b45778765-nh2m9\" (UID: \"97cfbecd-36ef-409b-94e9-f607a1fa2c42\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.920292 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8qsw\" (UniqueName: \"kubernetes.io/projected/6de35940-bef4-4dfa-9a83-08ba29d73399-kube-api-access-j8qsw\") pod \"cluster-image-registry-operator-dc59b4c8b-4dn69\" (UID: \"6de35940-bef4-4dfa-9a83-08ba29d73399\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4dn69" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.920390 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.920469 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d288ee23-1753-48f2-ab82-736defe5fe18-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9thpj\" (UID: \"d288ee23-1753-48f2-ab82-736defe5fe18\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.920545 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96-proxy-tls\") pod \"machine-config-controller-84d6567774-nf7jb\" (UID: \"2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nf7jb" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.920609 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc8e2d06-1cc2-4ea7-8d87-340d28740e20-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ch6wr\" (UID: \"fc8e2d06-1cc2-4ea7-8d87-340d28740e20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ch6wr" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.920676 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc8e2d06-1cc2-4ea7-8d87-340d28740e20-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ch6wr\" (UID: \"fc8e2d06-1cc2-4ea7-8d87-340d28740e20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ch6wr" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.920754 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fbfdd647-1d64-4d35-9af2-6dee52b4c860-serving-cert\") pod \"route-controller-manager-6576b87f9c-m2qxw\" (UID: \"fbfdd647-1d64-4d35-9af2-6dee52b4c860\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.920837 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e2a789d-6a90-4d60-881e-9562cd92e0a7-trusted-ca-bundle\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.920907 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fbcn\" (UniqueName: \"kubernetes.io/projected/42b66dc3-a385-4350-a943-50f062da35f7-kube-api-access-2fbcn\") pod \"downloads-7954f5f757-fn75b\" (UID: \"42b66dc3-a385-4350-a943-50f062da35f7\") " pod="openshift-console/downloads-7954f5f757-fn75b" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.920999 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99bab267-639b-48b1-abc4-8c0373200a39-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.921066 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.921144 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/97cfbecd-36ef-409b-94e9-f607a1fa2c42-etcd-client\") pod \"etcd-operator-b45778765-nh2m9\" (UID: \"97cfbecd-36ef-409b-94e9-f607a1fa2c42\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.921214 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.921282 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3c2d0611-58f8-4a7e-8280-361c80d62802-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xlczd\" (UID: \"3c2d0611-58f8-4a7e-8280-361c80d62802\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.921353 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0387af3d-8796-46b0-9282-9ecbda7fe3a7-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp5md\" (UID: \"0387af3d-8796-46b0-9282-9ecbda7fe3a7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp5md" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.921422 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bn272\" (UniqueName: \"kubernetes.io/projected/0e2a789d-6a90-4d60-881e-9562cd92e0a7-kube-api-access-bn272\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.920996 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c779f8ba-7614-49f1-be6d-a9e316ec59ba-audit-dir\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.921534 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eacb9f84-018a-4f64-b211-c9bedce50b9e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-vcj84\" (UID: \"eacb9f84-018a-4f64-b211-c9bedce50b9e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vcj84" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.921737 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c2d0611-58f8-4a7e-8280-361c80d62802-serving-cert\") pod \"controller-manager-879f6c89f-xlczd\" (UID: \"3c2d0611-58f8-4a7e-8280-361c80d62802\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.921862 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjjl8\" (UniqueName: \"kubernetes.io/projected/97cfbecd-36ef-409b-94e9-f607a1fa2c42-kube-api-access-rjjl8\") pod \"etcd-operator-b45778765-nh2m9\" (UID: \"97cfbecd-36ef-409b-94e9-f607a1fa2c42\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.922476 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c779f8ba-7614-49f1-be6d-a9e316ec59ba-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.922708 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/99bab267-639b-48b1-abc4-8c0373200a39-audit\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.922747 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.922771 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.922793 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c14fb55e-a42b-46c9-9521-6e8b60235166-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2lpzx\" (UID: \"c14fb55e-a42b-46c9-9521-6e8b60235166\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2lpzx" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.922812 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f3e2f962-69e3-4008-a45f-5c35677f7f36-machine-approver-tls\") pod \"machine-approver-56656f9798-xvrxj\" (UID: \"f3e2f962-69e3-4008-a45f-5c35677f7f36\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xvrxj" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.922830 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92-config\") pod \"machine-api-operator-5694c8668f-z4h55\" (UID: \"bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z4h55" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.922845 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnhkm\" (UniqueName: \"kubernetes.io/projected/c779f8ba-7614-49f1-be6d-a9e316ec59ba-kube-api-access-rnhkm\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.922862 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5b5da98c-0704-41c7-8563-707f7af93f41-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-99knh\" (UID: \"5b5da98c-0704-41c7-8563-707f7af93f41\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-99knh" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.922897 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/97cfbecd-36ef-409b-94e9-f607a1fa2c42-etcd-ca\") pod \"etcd-operator-b45778765-nh2m9\" (UID: \"97cfbecd-36ef-409b-94e9-f607a1fa2c42\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.922948 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e2a789d-6a90-4d60-881e-9562cd92e0a7-trusted-ca-bundle\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.923005 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvzb6\" (UniqueName: \"kubernetes.io/projected/d288ee23-1753-48f2-ab82-736defe5fe18-kube-api-access-tvzb6\") pod \"machine-config-operator-74547568cd-9thpj\" (UID: \"d288ee23-1753-48f2-ab82-736defe5fe18\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.923025 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-789j9\" (UniqueName: \"kubernetes.io/projected/fc8e2d06-1cc2-4ea7-8d87-340d28740e20-kube-api-access-789j9\") pod \"kube-storage-version-migrator-operator-b67b599dd-ch6wr\" (UID: \"fc8e2d06-1cc2-4ea7-8d87-340d28740e20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ch6wr" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.923047 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99bab267-639b-48b1-abc4-8c0373200a39-config\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.923065 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.923082 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92-images\") pod \"machine-api-operator-5694c8668f-z4h55\" (UID: \"bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z4h55" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.923100 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec-serving-cert\") pod \"console-operator-58897d9998-kvp9d\" (UID: \"ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec\") " pod="openshift-console-operator/console-operator-58897d9998-kvp9d" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.923116 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0387af3d-8796-46b0-9282-9ecbda7fe3a7-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp5md\" (UID: \"0387af3d-8796-46b0-9282-9ecbda7fe3a7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp5md" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.923133 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec-config\") pod \"console-operator-58897d9998-kvp9d\" (UID: \"ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec\") " pod="openshift-console-operator/console-operator-58897d9998-kvp9d" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.923162 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hv4x\" (UniqueName: \"kubernetes.io/projected/f3e2f962-69e3-4008-a45f-5c35677f7f36-kube-api-access-5hv4x\") pod \"machine-approver-56656f9798-xvrxj\" (UID: \"f3e2f962-69e3-4008-a45f-5c35677f7f36\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xvrxj" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.923179 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b25d77ec-57de-4c2a-b534-e98bf149b92a-serving-cert\") pod \"authentication-operator-69f744f599-jb6jw\" (UID: \"b25d77ec-57de-4c2a-b534-e98bf149b92a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jb6jw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.923199 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/99bab267-639b-48b1-abc4-8c0373200a39-audit-dir\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.923442 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/97cfbecd-36ef-409b-94e9-f607a1fa2c42-etcd-service-ca\") pod \"etcd-operator-b45778765-nh2m9\" (UID: \"97cfbecd-36ef-409b-94e9-f607a1fa2c42\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.923592 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/99bab267-639b-48b1-abc4-8c0373200a39-etcd-serving-ca\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.924095 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.924288 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3c2d0611-58f8-4a7e-8280-361c80d62802-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xlczd\" (UID: \"3c2d0611-58f8-4a7e-8280-361c80d62802\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.924483 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eacb9f84-018a-4f64-b211-c9bedce50b9e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-vcj84\" (UID: \"eacb9f84-018a-4f64-b211-c9bedce50b9e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vcj84" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.924599 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/99bab267-639b-48b1-abc4-8c0373200a39-audit\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.925696 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/99bab267-639b-48b1-abc4-8c0373200a39-audit-dir\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.925950 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2lpzx"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.926025 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97cfbecd-36ef-409b-94e9-f607a1fa2c42-config\") pod \"etcd-operator-b45778765-nh2m9\" (UID: \"97cfbecd-36ef-409b-94e9-f607a1fa2c42\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927136 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec-config\") pod \"console-operator-58897d9998-kvp9d\" (UID: \"ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec\") " pod="openshift-console-operator/console-operator-58897d9998-kvp9d" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927148 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99bab267-639b-48b1-abc4-8c0373200a39-config\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927252 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/6de35940-bef4-4dfa-9a83-08ba29d73399-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-4dn69\" (UID: \"6de35940-bef4-4dfa-9a83-08ba29d73399\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4dn69" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927293 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2rbv\" (UniqueName: \"kubernetes.io/projected/99bab267-639b-48b1-abc4-8c0373200a39-kube-api-access-f2rbv\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927315 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c14fb55e-a42b-46c9-9521-6e8b60235166-config\") pod \"kube-controller-manager-operator-78b949d7b-2lpzx\" (UID: \"c14fb55e-a42b-46c9-9521-6e8b60235166\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2lpzx" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927338 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0e2a789d-6a90-4d60-881e-9562cd92e0a7-console-config\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927375 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f3e2f962-69e3-4008-a45f-5c35677f7f36-auth-proxy-config\") pod \"machine-approver-56656f9798-xvrxj\" (UID: \"f3e2f962-69e3-4008-a45f-5c35677f7f36\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xvrxj" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927393 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b25d77ec-57de-4c2a-b534-e98bf149b92a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jb6jw\" (UID: \"b25d77ec-57de-4c2a-b534-e98bf149b92a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jb6jw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927408 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-z4h55\" (UID: \"bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z4h55" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927423 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knzbq\" (UniqueName: \"kubernetes.io/projected/8f71ba3e-c687-4ff7-9475-1e18ded764f6-kube-api-access-knzbq\") pod \"openshift-config-operator-7777fb866f-dwsm5\" (UID: \"8f71ba3e-c687-4ff7-9475-1e18ded764f6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927438 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c779f8ba-7614-49f1-be6d-a9e316ec59ba-serving-cert\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927452 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c2d0611-58f8-4a7e-8280-361c80d62802-config\") pod \"controller-manager-879f6c89f-xlczd\" (UID: \"3c2d0611-58f8-4a7e-8280-361c80d62802\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927467 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/832097a5-4691-42b6-99cc-38679071d5ee-audit-dir\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927483 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927508 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srhd7\" (UniqueName: \"kubernetes.io/projected/fbfdd647-1d64-4d35-9af2-6dee52b4c860-kube-api-access-srhd7\") pod \"route-controller-manager-6576b87f9c-m2qxw\" (UID: \"fbfdd647-1d64-4d35-9af2-6dee52b4c860\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927125 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927522 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99cl5\" (UniqueName: \"kubernetes.io/projected/b25d77ec-57de-4c2a-b534-e98bf149b92a-kube-api-access-99cl5\") pod \"authentication-operator-69f744f599-jb6jw\" (UID: \"b25d77ec-57de-4c2a-b534-e98bf149b92a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jb6jw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927539 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c779f8ba-7614-49f1-be6d-a9e316ec59ba-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927555 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/99bab267-639b-48b1-abc4-8c0373200a39-etcd-client\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927570 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-audit-policies\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927575 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92-images\") pod \"machine-api-operator-5694c8668f-z4h55\" (UID: \"bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z4h55" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927587 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927615 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3e2f962-69e3-4008-a45f-5c35677f7f36-config\") pod \"machine-approver-56656f9798-xvrxj\" (UID: \"f3e2f962-69e3-4008-a45f-5c35677f7f36\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xvrxj" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927629 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b25d77ec-57de-4c2a-b534-e98bf149b92a-service-ca-bundle\") pod \"authentication-operator-69f744f599-jb6jw\" (UID: \"b25d77ec-57de-4c2a-b534-e98bf149b92a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jb6jw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927645 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb7zc\" (UniqueName: \"kubernetes.io/projected/eacb9f84-018a-4f64-b211-c9bedce50b9e-kube-api-access-sb7zc\") pod \"openshift-apiserver-operator-796bbdcf4f-vcj84\" (UID: \"eacb9f84-018a-4f64-b211-c9bedce50b9e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vcj84" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927660 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/99bab267-639b-48b1-abc4-8c0373200a39-encryption-config\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927676 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927692 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c14fb55e-a42b-46c9-9521-6e8b60235166-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2lpzx\" (UID: \"c14fb55e-a42b-46c9-9521-6e8b60235166\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2lpzx" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927742 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/832097a5-4691-42b6-99cc-38679071d5ee-audit-dir\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927755 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbfdd647-1d64-4d35-9af2-6dee52b4c860-config\") pod \"route-controller-manager-6576b87f9c-m2qxw\" (UID: \"fbfdd647-1d64-4d35-9af2-6dee52b4c860\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927774 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b25d77ec-57de-4c2a-b534-e98bf149b92a-config\") pod \"authentication-operator-69f744f599-jb6jw\" (UID: \"b25d77ec-57de-4c2a-b534-e98bf149b92a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jb6jw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927797 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3c2d0611-58f8-4a7e-8280-361c80d62802-client-ca\") pod \"controller-manager-879f6c89f-xlczd\" (UID: \"3c2d0611-58f8-4a7e-8280-361c80d62802\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927815 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6de35940-bef4-4dfa-9a83-08ba29d73399-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-4dn69\" (UID: \"6de35940-bef4-4dfa-9a83-08ba29d73399\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4dn69" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927831 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/99bab267-639b-48b1-abc4-8c0373200a39-image-import-ca\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.927847 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8nmp\" (UniqueName: \"kubernetes.io/projected/832097a5-4691-42b6-99cc-38679071d5ee-kube-api-access-l8nmp\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.928654 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0e2a789d-6a90-4d60-881e-9562cd92e0a7-console-config\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.928682 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/99bab267-639b-48b1-abc4-8c0373200a39-node-pullsecrets\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.928703 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.928722 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fbfdd647-1d64-4d35-9af2-6dee52b4c860-client-ca\") pod \"route-controller-manager-6576b87f9c-m2qxw\" (UID: \"fbfdd647-1d64-4d35-9af2-6dee52b4c860\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.928736 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0e2a789d-6a90-4d60-881e-9562cd92e0a7-service-ca\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.928752 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eacb9f84-018a-4f64-b211-c9bedce50b9e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-vcj84\" (UID: \"eacb9f84-018a-4f64-b211-c9bedce50b9e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vcj84" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.928766 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99bab267-639b-48b1-abc4-8c0373200a39-serving-cert\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.928783 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0e2a789d-6a90-4d60-881e-9562cd92e0a7-console-serving-cert\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.929115 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97cfbecd-36ef-409b-94e9-f607a1fa2c42-config\") pod \"etcd-operator-b45778765-nh2m9\" (UID: \"97cfbecd-36ef-409b-94e9-f607a1fa2c42\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.929180 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f3e2f962-69e3-4008-a45f-5c35677f7f36-auth-proxy-config\") pod \"machine-approver-56656f9798-xvrxj\" (UID: \"f3e2f962-69e3-4008-a45f-5c35677f7f36\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xvrxj" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.930020 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c779f8ba-7614-49f1-be6d-a9e316ec59ba-serving-cert\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.930093 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b25d77ec-57de-4c2a-b534-e98bf149b92a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jb6jw\" (UID: \"b25d77ec-57de-4c2a-b534-e98bf149b92a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jb6jw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.930130 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0387af3d-8796-46b0-9282-9ecbda7fe3a7-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp5md\" (UID: \"0387af3d-8796-46b0-9282-9ecbda7fe3a7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp5md" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.930296 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.930378 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99bab267-639b-48b1-abc4-8c0373200a39-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.930741 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3e2f962-69e3-4008-a45f-5c35677f7f36-config\") pod \"machine-approver-56656f9798-xvrxj\" (UID: \"f3e2f962-69e3-4008-a45f-5c35677f7f36\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xvrxj" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.930928 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/97cfbecd-36ef-409b-94e9-f607a1fa2c42-etcd-client\") pod \"etcd-operator-b45778765-nh2m9\" (UID: \"97cfbecd-36ef-409b-94e9-f607a1fa2c42\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.930960 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5b5da98c-0704-41c7-8563-707f7af93f41-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-99knh\" (UID: \"5b5da98c-0704-41c7-8563-707f7af93f41\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-99knh" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.931152 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-audit-policies\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.931671 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.931918 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c2d0611-58f8-4a7e-8280-361c80d62802-serving-cert\") pod \"controller-manager-879f6c89f-xlczd\" (UID: \"3c2d0611-58f8-4a7e-8280-361c80d62802\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.932443 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.932457 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/99bab267-639b-48b1-abc4-8c0373200a39-node-pullsecrets\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.932620 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b25d77ec-57de-4c2a-b534-e98bf149b92a-service-ca-bundle\") pod \"authentication-operator-69f744f599-jb6jw\" (UID: \"b25d77ec-57de-4c2a-b534-e98bf149b92a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jb6jw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.933065 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b25d77ec-57de-4c2a-b534-e98bf149b92a-config\") pod \"authentication-operator-69f744f599-jb6jw\" (UID: \"b25d77ec-57de-4c2a-b534-e98bf149b92a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jb6jw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.933388 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/99bab267-639b-48b1-abc4-8c0373200a39-image-import-ca\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.933429 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0e2a789d-6a90-4d60-881e-9562cd92e0a7-service-ca\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.934073 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/99bab267-639b-48b1-abc4-8c0373200a39-etcd-client\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.935523 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3c2d0611-58f8-4a7e-8280-361c80d62802-client-ca\") pod \"controller-manager-879f6c89f-xlczd\" (UID: \"3c2d0611-58f8-4a7e-8280-361c80d62802\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.935577 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c2d0611-58f8-4a7e-8280-361c80d62802-config\") pod \"controller-manager-879f6c89f-xlczd\" (UID: \"3c2d0611-58f8-4a7e-8280-361c80d62802\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.935910 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c779f8ba-7614-49f1-be6d-a9e316ec59ba-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.936071 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fbfdd647-1d64-4d35-9af2-6dee52b4c860-client-ca\") pod \"route-controller-manager-6576b87f9c-m2qxw\" (UID: \"fbfdd647-1d64-4d35-9af2-6dee52b4c860\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.936431 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.936982 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99bab267-639b-48b1-abc4-8c0373200a39-serving-cert\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.937047 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d288ee23-1753-48f2-ab82-736defe5fe18-images\") pod \"machine-config-operator-74547568cd-9thpj\" (UID: \"d288ee23-1753-48f2-ab82-736defe5fe18\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.937073 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/554abf87-b1ba-45b1-8130-95b40da3b8bf-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mrprd\" (UID: \"554abf87-b1ba-45b1-8130-95b40da3b8bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mrprd" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.937098 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec-trusted-ca\") pod \"console-operator-58897d9998-kvp9d\" (UID: \"ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec\") " pod="openshift-console-operator/console-operator-58897d9998-kvp9d" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.937132 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8tm6\" (UniqueName: \"kubernetes.io/projected/bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92-kube-api-access-k8tm6\") pod \"machine-api-operator-5694c8668f-z4h55\" (UID: \"bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z4h55" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.937154 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6de35940-bef4-4dfa-9a83-08ba29d73399-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-4dn69\" (UID: \"6de35940-bef4-4dfa-9a83-08ba29d73399\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4dn69" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.937207 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6de35940-bef4-4dfa-9a83-08ba29d73399-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-4dn69\" (UID: \"6de35940-bef4-4dfa-9a83-08ba29d73399\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4dn69" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.937227 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8f71ba3e-c687-4ff7-9475-1e18ded764f6-available-featuregates\") pod \"openshift-config-operator-7777fb866f-dwsm5\" (UID: \"8f71ba3e-c687-4ff7-9475-1e18ded764f6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.937264 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmd5b\" (UniqueName: \"kubernetes.io/projected/5b5da98c-0704-41c7-8563-707f7af93f41-kube-api-access-hmd5b\") pod \"cluster-samples-operator-665b6dd947-99knh\" (UID: \"5b5da98c-0704-41c7-8563-707f7af93f41\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-99knh" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.937284 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d288ee23-1753-48f2-ab82-736defe5fe18-proxy-tls\") pod \"machine-config-operator-74547568cd-9thpj\" (UID: \"d288ee23-1753-48f2-ab82-736defe5fe18\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.937305 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c779f8ba-7614-49f1-be6d-a9e316ec59ba-audit-policies\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.937345 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.937367 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0e2a789d-6a90-4d60-881e-9562cd92e0a7-oauth-serving-cert\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.937385 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f71ba3e-c687-4ff7-9475-1e18ded764f6-serving-cert\") pod \"openshift-config-operator-7777fb866f-dwsm5\" (UID: \"8f71ba3e-c687-4ff7-9475-1e18ded764f6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.937418 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c779f8ba-7614-49f1-be6d-a9e316ec59ba-etcd-client\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.937437 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/554abf87-b1ba-45b1-8130-95b40da3b8bf-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mrprd\" (UID: \"554abf87-b1ba-45b1-8130-95b40da3b8bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mrprd" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.938010 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c779f8ba-7614-49f1-be6d-a9e316ec59ba-audit-policies\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.938008 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.938016 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec-trusted-ca\") pod \"console-operator-58897d9998-kvp9d\" (UID: \"ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec\") " pod="openshift-console-operator/console-operator-58897d9998-kvp9d" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.938530 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0e2a789d-6a90-4d60-881e-9562cd92e0a7-oauth-serving-cert\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.938569 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-z4h55\" (UID: \"bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z4h55" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.938645 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/6de35940-bef4-4dfa-9a83-08ba29d73399-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-4dn69\" (UID: \"6de35940-bef4-4dfa-9a83-08ba29d73399\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4dn69" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.939229 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.939369 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8f71ba3e-c687-4ff7-9475-1e18ded764f6-available-featuregates\") pod \"openshift-config-operator-7777fb866f-dwsm5\" (UID: \"8f71ba3e-c687-4ff7-9475-1e18ded764f6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.939567 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.939733 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0e2a789d-6a90-4d60-881e-9562cd92e0a7-console-serving-cert\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.940161 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.940173 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eacb9f84-018a-4f64-b211-c9bedce50b9e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-vcj84\" (UID: \"eacb9f84-018a-4f64-b211-c9bedce50b9e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vcj84" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.940447 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c779f8ba-7614-49f1-be6d-a9e316ec59ba-etcd-client\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.940473 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92-config\") pod \"machine-api-operator-5694c8668f-z4h55\" (UID: \"bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z4h55" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.944801 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.945746 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f3e2f962-69e3-4008-a45f-5c35677f7f36-machine-approver-tls\") pod \"machine-approver-56656f9798-xvrxj\" (UID: \"f3e2f962-69e3-4008-a45f-5c35677f7f36\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xvrxj" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.945830 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-nf7jb"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.945846 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f71ba3e-c687-4ff7-9475-1e18ded764f6-serving-cert\") pod \"openshift-config-operator-7777fb866f-dwsm5\" (UID: \"8f71ba3e-c687-4ff7-9475-1e18ded764f6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.946411 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vcj84"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.947489 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.948558 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.950915 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/99bab267-639b-48b1-abc4-8c0373200a39-encryption-config\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.951709 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbfdd647-1d64-4d35-9af2-6dee52b4c860-config\") pod \"route-controller-manager-6576b87f9c-m2qxw\" (UID: \"fbfdd647-1d64-4d35-9af2-6dee52b4c860\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.952601 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.954336 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.954610 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6jl75"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.956556 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gnjmm"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.958114 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-kvp9d"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.958126 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fbfdd647-1d64-4d35-9af2-6dee52b4c860-serving-cert\") pod \"route-controller-manager-6576b87f9c-m2qxw\" (UID: \"fbfdd647-1d64-4d35-9af2-6dee52b4c860\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.959413 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.960254 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97cfbecd-36ef-409b-94e9-f607a1fa2c42-serving-cert\") pod \"etcd-operator-b45778765-nh2m9\" (UID: \"97cfbecd-36ef-409b-94e9-f607a1fa2c42\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.960376 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c779f8ba-7614-49f1-be6d-a9e316ec59ba-encryption-config\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.960389 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.960396 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0387af3d-8796-46b0-9282-9ecbda7fe3a7-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp5md\" (UID: \"0387af3d-8796-46b0-9282-9ecbda7fe3a7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp5md" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.960461 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0e2a789d-6a90-4d60-881e-9562cd92e0a7-console-oauth-config\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.960873 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b25d77ec-57de-4c2a-b534-e98bf149b92a-serving-cert\") pod \"authentication-operator-69f744f599-jb6jw\" (UID: \"b25d77ec-57de-4c2a-b534-e98bf149b92a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jb6jw" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.962157 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-sv7xw"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.963686 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec-serving-cert\") pod \"console-operator-58897d9998-kvp9d\" (UID: \"ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec\") " pod="openshift-console-operator/console-operator-58897d9998-kvp9d" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.964044 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.965968 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-nh2m9"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.967524 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp5md"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.968402 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.969094 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.970094 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-l2t56"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.971057 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-vhtdt"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.971804 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-vhtdt" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.972021 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-4nghl"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.973022 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chzp2"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.973033 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-4nghl" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.974021 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-44gfk"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.974952 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mrprd"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.975891 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-zkbcz"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.976886 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-z4h55"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.978095 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-99knh"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.979390 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xtzbx"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.981191 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-h8b4r"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.982227 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jzg2m"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.983318 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-jcdhl"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.984713 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-4nghl"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.985626 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-kfqcf"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.986555 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.987612 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pbrmk"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.988345 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.989007 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ljnh7"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.990011 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-vhtdt"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.991108 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sq9mx"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.998578 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-zxb4v"] Jan 29 16:12:13 crc kubenswrapper[4714]: I0129 16:12:13.999435 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-zxb4v" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.008143 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.028284 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.039020 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c14fb55e-a42b-46c9-9521-6e8b60235166-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2lpzx\" (UID: \"c14fb55e-a42b-46c9-9521-6e8b60235166\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2lpzx" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.039066 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-789j9\" (UniqueName: \"kubernetes.io/projected/fc8e2d06-1cc2-4ea7-8d87-340d28740e20-kube-api-access-789j9\") pod \"kube-storage-version-migrator-operator-b67b599dd-ch6wr\" (UID: \"fc8e2d06-1cc2-4ea7-8d87-340d28740e20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ch6wr" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.039105 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvzb6\" (UniqueName: \"kubernetes.io/projected/d288ee23-1753-48f2-ab82-736defe5fe18-kube-api-access-tvzb6\") pod \"machine-config-operator-74547568cd-9thpj\" (UID: \"d288ee23-1753-48f2-ab82-736defe5fe18\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.039158 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c14fb55e-a42b-46c9-9521-6e8b60235166-config\") pod \"kube-controller-manager-operator-78b949d7b-2lpzx\" (UID: \"c14fb55e-a42b-46c9-9521-6e8b60235166\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2lpzx" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.039269 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c14fb55e-a42b-46c9-9521-6e8b60235166-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2lpzx\" (UID: \"c14fb55e-a42b-46c9-9521-6e8b60235166\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2lpzx" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.039317 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/554abf87-b1ba-45b1-8130-95b40da3b8bf-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mrprd\" (UID: \"554abf87-b1ba-45b1-8130-95b40da3b8bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mrprd" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.039343 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d288ee23-1753-48f2-ab82-736defe5fe18-images\") pod \"machine-config-operator-74547568cd-9thpj\" (UID: \"d288ee23-1753-48f2-ab82-736defe5fe18\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.039376 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d288ee23-1753-48f2-ab82-736defe5fe18-proxy-tls\") pod \"machine-config-operator-74547568cd-9thpj\" (UID: \"d288ee23-1753-48f2-ab82-736defe5fe18\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.039421 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/554abf87-b1ba-45b1-8130-95b40da3b8bf-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mrprd\" (UID: \"554abf87-b1ba-45b1-8130-95b40da3b8bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mrprd" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.039454 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45cdw\" (UniqueName: \"kubernetes.io/projected/2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96-kube-api-access-45cdw\") pod \"machine-config-controller-84d6567774-nf7jb\" (UID: \"2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nf7jb" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.039491 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-nf7jb\" (UID: \"2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nf7jb" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.039508 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwvc4\" (UniqueName: \"kubernetes.io/projected/554abf87-b1ba-45b1-8130-95b40da3b8bf-kube-api-access-zwvc4\") pod \"openshift-controller-manager-operator-756b6f6bc6-mrprd\" (UID: \"554abf87-b1ba-45b1-8130-95b40da3b8bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mrprd" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.039524 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96-proxy-tls\") pod \"machine-config-controller-84d6567774-nf7jb\" (UID: \"2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nf7jb" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.039539 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc8e2d06-1cc2-4ea7-8d87-340d28740e20-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ch6wr\" (UID: \"fc8e2d06-1cc2-4ea7-8d87-340d28740e20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ch6wr" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.039563 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d288ee23-1753-48f2-ab82-736defe5fe18-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9thpj\" (UID: \"d288ee23-1753-48f2-ab82-736defe5fe18\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.039580 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc8e2d06-1cc2-4ea7-8d87-340d28740e20-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ch6wr\" (UID: \"fc8e2d06-1cc2-4ea7-8d87-340d28740e20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ch6wr" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.040794 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-nf7jb\" (UID: \"2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nf7jb" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.040956 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d288ee23-1753-48f2-ab82-736defe5fe18-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9thpj\" (UID: \"d288ee23-1753-48f2-ab82-736defe5fe18\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.042896 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96-proxy-tls\") pod \"machine-config-controller-84d6567774-nf7jb\" (UID: \"2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nf7jb" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.048158 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.068974 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.088402 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.109142 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.128439 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.149248 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.168558 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.189245 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.208577 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.228795 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.249273 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.253795 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d288ee23-1753-48f2-ab82-736defe5fe18-proxy-tls\") pod \"machine-config-operator-74547568cd-9thpj\" (UID: \"d288ee23-1753-48f2-ab82-736defe5fe18\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.268769 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.271360 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d288ee23-1753-48f2-ab82-736defe5fe18-images\") pod \"machine-config-operator-74547568cd-9thpj\" (UID: \"d288ee23-1753-48f2-ab82-736defe5fe18\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.288045 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.308324 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.328848 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.334252 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/554abf87-b1ba-45b1-8130-95b40da3b8bf-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mrprd\" (UID: \"554abf87-b1ba-45b1-8130-95b40da3b8bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mrprd" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.349455 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.350755 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/554abf87-b1ba-45b1-8130-95b40da3b8bf-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mrprd\" (UID: \"554abf87-b1ba-45b1-8130-95b40da3b8bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mrprd" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.369563 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.388392 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.407816 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.428556 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.448551 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.469576 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.487796 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.516882 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.530415 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.549121 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.568979 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.589139 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.593415 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c14fb55e-a42b-46c9-9521-6e8b60235166-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2lpzx\" (UID: \"c14fb55e-a42b-46c9-9521-6e8b60235166\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2lpzx" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.609886 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.610644 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c14fb55e-a42b-46c9-9521-6e8b60235166-config\") pod \"kube-controller-manager-operator-78b949d7b-2lpzx\" (UID: \"c14fb55e-a42b-46c9-9521-6e8b60235166\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2lpzx" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.648711 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.668873 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.688897 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.715195 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.729708 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.748819 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.754664 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc8e2d06-1cc2-4ea7-8d87-340d28740e20-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ch6wr\" (UID: \"fc8e2d06-1cc2-4ea7-8d87-340d28740e20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ch6wr" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.769985 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.788103 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.809663 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.828608 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.848197 4714 request.go:700] Waited for 1.000785101s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/configmaps?fieldSelector=metadata.name%3Dconfig&limit=500&resourceVersion=0 Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.850571 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.862200 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc8e2d06-1cc2-4ea7-8d87-340d28740e20-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ch6wr\" (UID: \"fc8e2d06-1cc2-4ea7-8d87-340d28740e20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ch6wr" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.868328 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.888971 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.908542 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.929553 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.949492 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.968741 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 29 16:12:14 crc kubenswrapper[4714]: I0129 16:12:14.988540 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.028352 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.048843 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.068592 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.088626 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.109486 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.128841 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.148733 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.169332 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.188844 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.209563 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.229015 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.248853 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.268798 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.288732 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.308475 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.328849 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.348572 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.368688 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.387777 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.408985 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.428830 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.450155 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.469033 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.489230 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.509157 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.529109 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.571142 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x5xk\" (UniqueName: \"kubernetes.io/projected/ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec-kube-api-access-4x5xk\") pod \"console-operator-58897d9998-kvp9d\" (UID: \"ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec\") " pod="openshift-console-operator/console-operator-58897d9998-kvp9d" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.597622 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fbcn\" (UniqueName: \"kubernetes.io/projected/42b66dc3-a385-4350-a943-50f062da35f7-kube-api-access-2fbcn\") pod \"downloads-7954f5f757-fn75b\" (UID: \"42b66dc3-a385-4350-a943-50f062da35f7\") " pod="openshift-console/downloads-7954f5f757-fn75b" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.607224 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-kvp9d" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.619689 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnhkm\" (UniqueName: \"kubernetes.io/projected/c779f8ba-7614-49f1-be6d-a9e316ec59ba-kube-api-access-rnhkm\") pod \"apiserver-7bbb656c7d-kgl5s\" (UID: \"c779f8ba-7614-49f1-be6d-a9e316ec59ba\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.635726 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjjl8\" (UniqueName: \"kubernetes.io/projected/97cfbecd-36ef-409b-94e9-f607a1fa2c42-kube-api-access-rjjl8\") pod \"etcd-operator-b45778765-nh2m9\" (UID: \"97cfbecd-36ef-409b-94e9-f607a1fa2c42\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.647584 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0387af3d-8796-46b0-9282-9ecbda7fe3a7-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp5md\" (UID: \"0387af3d-8796-46b0-9282-9ecbda7fe3a7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp5md" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.666771 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bn272\" (UniqueName: \"kubernetes.io/projected/0e2a789d-6a90-4d60-881e-9562cd92e0a7-kube-api-access-bn272\") pod \"console-f9d7485db-m2g9h\" (UID: \"0e2a789d-6a90-4d60-881e-9562cd92e0a7\") " pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.680699 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.686467 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hv4x\" (UniqueName: \"kubernetes.io/projected/f3e2f962-69e3-4008-a45f-5c35677f7f36-kube-api-access-5hv4x\") pod \"machine-approver-56656f9798-xvrxj\" (UID: \"f3e2f962-69e3-4008-a45f-5c35677f7f36\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xvrxj" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.707746 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8qsw\" (UniqueName: \"kubernetes.io/projected/6de35940-bef4-4dfa-9a83-08ba29d73399-kube-api-access-j8qsw\") pod \"cluster-image-registry-operator-dc59b4c8b-4dn69\" (UID: \"6de35940-bef4-4dfa-9a83-08ba29d73399\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4dn69" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.716399 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.724879 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk8rh\" (UniqueName: \"kubernetes.io/projected/3c2d0611-58f8-4a7e-8280-361c80d62802-kube-api-access-pk8rh\") pod \"controller-manager-879f6c89f-xlczd\" (UID: \"3c2d0611-58f8-4a7e-8280-361c80d62802\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.750223 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knzbq\" (UniqueName: \"kubernetes.io/projected/8f71ba3e-c687-4ff7-9475-1e18ded764f6-kube-api-access-knzbq\") pod \"openshift-config-operator-7777fb866f-dwsm5\" (UID: \"8f71ba3e-c687-4ff7-9475-1e18ded764f6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.765247 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srhd7\" (UniqueName: \"kubernetes.io/projected/fbfdd647-1d64-4d35-9af2-6dee52b4c860-kube-api-access-srhd7\") pod \"route-controller-manager-6576b87f9c-m2qxw\" (UID: \"fbfdd647-1d64-4d35-9af2-6dee52b4c860\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.777292 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.786608 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2rbv\" (UniqueName: \"kubernetes.io/projected/99bab267-639b-48b1-abc4-8c0373200a39-kube-api-access-f2rbv\") pod \"apiserver-76f77b778f-6jl75\" (UID: \"99bab267-639b-48b1-abc4-8c0373200a39\") " pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.794037 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-fn75b" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.809091 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.811157 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99cl5\" (UniqueName: \"kubernetes.io/projected/b25d77ec-57de-4c2a-b534-e98bf149b92a-kube-api-access-99cl5\") pod \"authentication-operator-69f744f599-jb6jw\" (UID: \"b25d77ec-57de-4c2a-b534-e98bf149b92a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jb6jw" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.814710 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp5md" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.828896 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb7zc\" (UniqueName: \"kubernetes.io/projected/eacb9f84-018a-4f64-b211-c9bedce50b9e-kube-api-access-sb7zc\") pod \"openshift-apiserver-operator-796bbdcf4f-vcj84\" (UID: \"eacb9f84-018a-4f64-b211-c9bedce50b9e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vcj84" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.835743 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xvrxj" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.844741 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-jb6jw" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.846297 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8nmp\" (UniqueName: \"kubernetes.io/projected/832097a5-4691-42b6-99cc-38679071d5ee-kube-api-access-l8nmp\") pod \"oauth-openshift-558db77b4-h8b4r\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.847551 4714 request.go:700] Waited for 1.910213618s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-api/serviceaccounts/machine-api-operator/token Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.866969 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8tm6\" (UniqueName: \"kubernetes.io/projected/bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92-kube-api-access-k8tm6\") pod \"machine-api-operator-5694c8668f-z4h55\" (UID: \"bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z4h55" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.871054 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.898443 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmd5b\" (UniqueName: \"kubernetes.io/projected/5b5da98c-0704-41c7-8563-707f7af93f41-kube-api-access-hmd5b\") pod \"cluster-samples-operator-665b6dd947-99knh\" (UID: \"5b5da98c-0704-41c7-8563-707f7af93f41\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-99knh" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.906205 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6de35940-bef4-4dfa-9a83-08ba29d73399-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-4dn69\" (UID: \"6de35940-bef4-4dfa-9a83-08ba29d73399\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4dn69" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.909748 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.916429 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-kvp9d"] Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.916691 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.930102 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.934044 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s"] Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.947589 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xvrxj" event={"ID":"f3e2f962-69e3-4008-a45f-5c35677f7f36","Type":"ContainerStarted","Data":"e6417ed16b50250c1230123b28ea7f944dd6f6be42cb84b33ddc22f04045662b"} Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.948544 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.953137 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vcj84" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.968817 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.987826 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-m2g9h"] Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.990259 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:15 crc kubenswrapper[4714]: I0129 16:12:15.991501 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.005830 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-z4h55" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.013621 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.023300 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4dn69" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.029017 4714 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.051836 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.070122 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.085042 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-99knh" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.089718 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.100697 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.128003 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-789j9\" (UniqueName: \"kubernetes.io/projected/fc8e2d06-1cc2-4ea7-8d87-340d28740e20-kube-api-access-789j9\") pod \"kube-storage-version-migrator-operator-b67b599dd-ch6wr\" (UID: \"fc8e2d06-1cc2-4ea7-8d87-340d28740e20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ch6wr" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.154647 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvzb6\" (UniqueName: \"kubernetes.io/projected/d288ee23-1753-48f2-ab82-736defe5fe18-kube-api-access-tvzb6\") pod \"machine-config-operator-74547568cd-9thpj\" (UID: \"d288ee23-1753-48f2-ab82-736defe5fe18\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.167028 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c14fb55e-a42b-46c9-9521-6e8b60235166-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2lpzx\" (UID: \"c14fb55e-a42b-46c9-9521-6e8b60235166\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2lpzx" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.181752 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2lpzx" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.188822 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwvc4\" (UniqueName: \"kubernetes.io/projected/554abf87-b1ba-45b1-8130-95b40da3b8bf-kube-api-access-zwvc4\") pod \"openshift-controller-manager-operator-756b6f6bc6-mrprd\" (UID: \"554abf87-b1ba-45b1-8130-95b40da3b8bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mrprd" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.195828 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jb6jw"] Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.198035 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ch6wr" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.211242 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45cdw\" (UniqueName: \"kubernetes.io/projected/2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96-kube-api-access-45cdw\") pod \"machine-config-controller-84d6567774-nf7jb\" (UID: \"2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nf7jb" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.216977 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.217032 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xlczd"] Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.252095 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vcj84"] Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.262728 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mrprd" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.272259 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfkhr\" (UniqueName: \"kubernetes.io/projected/1d9869e2-6f55-4246-8ed0-b8af9dab3f74-kube-api-access-gfkhr\") pod \"ingress-operator-5b745b69d9-r6cxt\" (UID: \"1d9869e2-6f55-4246-8ed0-b8af9dab3f74\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.272311 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80515d06-c09e-4c9d-a90f-43cc84edf4c9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-l2t56\" (UID: \"80515d06-c09e-4c9d-a90f-43cc84edf4c9\") " pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.272399 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfc66\" (UniqueName: \"kubernetes.io/projected/80515d06-c09e-4c9d-a90f-43cc84edf4c9-kube-api-access-xfc66\") pod \"marketplace-operator-79b997595-l2t56\" (UID: \"80515d06-c09e-4c9d-a90f-43cc84edf4c9\") " pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.272526 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfz7c\" (UniqueName: \"kubernetes.io/projected/cc84f60e-094e-4924-b6f1-f0a8ab81aa4e-kube-api-access-zfz7c\") pod \"catalog-operator-68c6474976-jzg2m\" (UID: \"cc84f60e-094e-4924-b6f1-f0a8ab81aa4e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jzg2m" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.272588 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/48be8ad8-4c02-4bea-a143-449763b39d54-trusted-ca\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.272626 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/48be8ad8-4c02-4bea-a143-449763b39d54-registry-tls\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.272685 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7a1dfb55-8680-4cbe-bd78-caca2e847caf-stats-auth\") pod \"router-default-5444994796-lz6mw\" (UID: \"7a1dfb55-8680-4cbe-bd78-caca2e847caf\") " pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.272710 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/48be8ad8-4c02-4bea-a143-449763b39d54-bound-sa-token\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.272777 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjnvs\" (UniqueName: \"kubernetes.io/projected/b7cf219f-4e80-47fc-b349-ea5c7eab6d9d-kube-api-access-wjnvs\") pod \"dns-operator-744455d44c-sv7xw\" (UID: \"b7cf219f-4e80-47fc-b349-ea5c7eab6d9d\") " pod="openshift-dns-operator/dns-operator-744455d44c-sv7xw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.272869 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/48be8ad8-4c02-4bea-a143-449763b39d54-registry-certificates\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.272992 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cc84f60e-094e-4924-b6f1-f0a8ab81aa4e-srv-cert\") pod \"catalog-operator-68c6474976-jzg2m\" (UID: \"cc84f60e-094e-4924-b6f1-f0a8ab81aa4e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jzg2m" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.273024 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6f4x\" (UniqueName: \"kubernetes.io/projected/7a1dfb55-8680-4cbe-bd78-caca2e847caf-kube-api-access-k6f4x\") pod \"router-default-5444994796-lz6mw\" (UID: \"7a1dfb55-8680-4cbe-bd78-caca2e847caf\") " pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.273068 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1d9869e2-6f55-4246-8ed0-b8af9dab3f74-bound-sa-token\") pod \"ingress-operator-5b745b69d9-r6cxt\" (UID: \"1d9869e2-6f55-4246-8ed0-b8af9dab3f74\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.273093 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a1dfb55-8680-4cbe-bd78-caca2e847caf-service-ca-bundle\") pod \"router-default-5444994796-lz6mw\" (UID: \"7a1dfb55-8680-4cbe-bd78-caca2e847caf\") " pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.273320 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cc84f60e-094e-4924-b6f1-f0a8ab81aa4e-profile-collector-cert\") pod \"catalog-operator-68c6474976-jzg2m\" (UID: \"cc84f60e-094e-4924-b6f1-f0a8ab81aa4e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jzg2m" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.273356 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n67sj\" (UniqueName: \"kubernetes.io/projected/2fa1ede8-3ea3-421d-929d-f6bf9cc1db0e-kube-api-access-n67sj\") pod \"migrator-59844c95c7-xtzbx\" (UID: \"2fa1ede8-3ea3-421d-929d-f6bf9cc1db0e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xtzbx" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.273396 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1d9869e2-6f55-4246-8ed0-b8af9dab3f74-metrics-tls\") pod \"ingress-operator-5b745b69d9-r6cxt\" (UID: \"1d9869e2-6f55-4246-8ed0-b8af9dab3f74\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.273462 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a1dfb55-8680-4cbe-bd78-caca2e847caf-metrics-certs\") pod \"router-default-5444994796-lz6mw\" (UID: \"7a1dfb55-8680-4cbe-bd78-caca2e847caf\") " pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.273485 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1d9869e2-6f55-4246-8ed0-b8af9dab3f74-trusted-ca\") pod \"ingress-operator-5b745b69d9-r6cxt\" (UID: \"1d9869e2-6f55-4246-8ed0-b8af9dab3f74\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.273596 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/80515d06-c09e-4c9d-a90f-43cc84edf4c9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-l2t56\" (UID: \"80515d06-c09e-4c9d-a90f-43cc84edf4c9\") " pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.273627 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh5km\" (UniqueName: \"kubernetes.io/projected/48be8ad8-4c02-4bea-a143-449763b39d54-kube-api-access-wh5km\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.273661 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/48be8ad8-4c02-4bea-a143-449763b39d54-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.273686 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/48be8ad8-4c02-4bea-a143-449763b39d54-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.273744 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.273816 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b7cf219f-4e80-47fc-b349-ea5c7eab6d9d-metrics-tls\") pod \"dns-operator-744455d44c-sv7xw\" (UID: \"b7cf219f-4e80-47fc-b349-ea5c7eab6d9d\") " pod="openshift-dns-operator/dns-operator-744455d44c-sv7xw" Jan 29 16:12:16 crc kubenswrapper[4714]: E0129 16:12:16.274595 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:16.774581357 +0000 UTC m=+143.295082467 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.275545 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7a1dfb55-8680-4cbe-bd78-caca2e847caf-default-certificate\") pod \"router-default-5444994796-lz6mw\" (UID: \"7a1dfb55-8680-4cbe-bd78-caca2e847caf\") " pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.329585 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-fn75b"] Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.329674 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5"] Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.333076 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-nh2m9"] Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.341275 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp5md"] Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395111 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395329 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfz7c\" (UniqueName: \"kubernetes.io/projected/cc84f60e-094e-4924-b6f1-f0a8ab81aa4e-kube-api-access-zfz7c\") pod \"catalog-operator-68c6474976-jzg2m\" (UID: \"cc84f60e-094e-4924-b6f1-f0a8ab81aa4e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jzg2m" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395402 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b632d84-c711-419a-9e24-bdb4c6e9aef6-config\") pod \"kube-apiserver-operator-766d6c64bb-chzp2\" (UID: \"1b632d84-c711-419a-9e24-bdb4c6e9aef6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chzp2" Jan 29 16:12:16 crc kubenswrapper[4714]: E0129 16:12:16.395417 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:16.895398652 +0000 UTC m=+143.415899772 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395456 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/38f843fe-c20b-4bc7-8f45-4ccd4b7be5a5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-ljnh7\" (UID: \"38f843fe-c20b-4bc7-8f45-4ccd4b7be5a5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ljnh7" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395473 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45vkr\" (UniqueName: \"kubernetes.io/projected/fcaee576-dff0-4a67-a0b1-7347b3030729-kube-api-access-45vkr\") pod \"dns-default-jcdhl\" (UID: \"fcaee576-dff0-4a67-a0b1-7347b3030729\") " pod="openshift-dns/dns-default-jcdhl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395535 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/714cef39-2960-4a25-ac81-a4e65a115eb3-plugins-dir\") pod \"csi-hostpathplugin-4nghl\" (UID: \"714cef39-2960-4a25-ac81-a4e65a115eb3\") " pod="hostpath-provisioner/csi-hostpathplugin-4nghl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.394782 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-h8b4r"] Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395561 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/48be8ad8-4c02-4bea-a143-449763b39d54-registry-tls\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395580 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/48be8ad8-4c02-4bea-a143-449763b39d54-trusted-ca\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395599 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fzjx\" (UniqueName: \"kubernetes.io/projected/cb979c55-3027-4d92-94b9-cd17c32e6331-kube-api-access-7fzjx\") pod \"collect-profiles-29495040-5mkf8\" (UID: \"cb979c55-3027-4d92-94b9-cd17c32e6331\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395625 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqcjl\" (UniqueName: \"kubernetes.io/projected/38f843fe-c20b-4bc7-8f45-4ccd4b7be5a5-kube-api-access-fqcjl\") pod \"olm-operator-6b444d44fb-ljnh7\" (UID: \"38f843fe-c20b-4bc7-8f45-4ccd4b7be5a5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ljnh7" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395640 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb979c55-3027-4d92-94b9-cd17c32e6331-secret-volume\") pod \"collect-profiles-29495040-5mkf8\" (UID: \"cb979c55-3027-4d92-94b9-cd17c32e6331\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395655 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf5f9\" (UniqueName: \"kubernetes.io/projected/f18250a8-66c1-445d-9452-081de13b24f7-kube-api-access-lf5f9\") pod \"multus-admission-controller-857f4d67dd-zkbcz\" (UID: \"f18250a8-66c1-445d-9452-081de13b24f7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zkbcz" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395673 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfkhq\" (UniqueName: \"kubernetes.io/projected/1fd5b799-74c2-4ffa-b3d9-6745c66ba28f-kube-api-access-vfkhq\") pod \"machine-config-server-zxb4v\" (UID: \"1fd5b799-74c2-4ffa-b3d9-6745c66ba28f\") " pod="openshift-machine-config-operator/machine-config-server-zxb4v" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395690 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7a1dfb55-8680-4cbe-bd78-caca2e847caf-stats-auth\") pod \"router-default-5444994796-lz6mw\" (UID: \"7a1dfb55-8680-4cbe-bd78-caca2e847caf\") " pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395713 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/48be8ad8-4c02-4bea-a143-449763b39d54-bound-sa-token\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395741 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjnvs\" (UniqueName: \"kubernetes.io/projected/b7cf219f-4e80-47fc-b349-ea5c7eab6d9d-kube-api-access-wjnvs\") pod \"dns-operator-744455d44c-sv7xw\" (UID: \"b7cf219f-4e80-47fc-b349-ea5c7eab6d9d\") " pod="openshift-dns-operator/dns-operator-744455d44c-sv7xw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395757 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab-tmpfs\") pod \"packageserver-d55dfcdfc-v68nn\" (UID: \"2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395774 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6k2j\" (UniqueName: \"kubernetes.io/projected/8062d225-aa57-48df-bf28-2254ecc4f635-kube-api-access-z6k2j\") pod \"control-plane-machine-set-operator-78cbb6b69f-sq9mx\" (UID: \"8062d225-aa57-48df-bf28-2254ecc4f635\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sq9mx" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395797 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1fd5b799-74c2-4ffa-b3d9-6745c66ba28f-node-bootstrap-token\") pod \"machine-config-server-zxb4v\" (UID: \"1fd5b799-74c2-4ffa-b3d9-6745c66ba28f\") " pod="openshift-machine-config-operator/machine-config-server-zxb4v" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395811 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p4cl\" (UniqueName: \"kubernetes.io/projected/2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab-kube-api-access-4p4cl\") pod \"packageserver-d55dfcdfc-v68nn\" (UID: \"2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395824 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwsfq\" (UniqueName: \"kubernetes.io/projected/714cef39-2960-4a25-ac81-a4e65a115eb3-kube-api-access-xwsfq\") pod \"csi-hostpathplugin-4nghl\" (UID: \"714cef39-2960-4a25-ac81-a4e65a115eb3\") " pod="hostpath-provisioner/csi-hostpathplugin-4nghl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395843 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8062d225-aa57-48df-bf28-2254ecc4f635-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sq9mx\" (UID: \"8062d225-aa57-48df-bf28-2254ecc4f635\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sq9mx" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.395876 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/48be8ad8-4c02-4bea-a143-449763b39d54-registry-certificates\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.396004 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb979c55-3027-4d92-94b9-cd17c32e6331-config-volume\") pod \"collect-profiles-29495040-5mkf8\" (UID: \"cb979c55-3027-4d92-94b9-cd17c32e6331\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.396450 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cc84f60e-094e-4924-b6f1-f0a8ab81aa4e-srv-cert\") pod \"catalog-operator-68c6474976-jzg2m\" (UID: \"cc84f60e-094e-4924-b6f1-f0a8ab81aa4e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jzg2m" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.396498 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b632d84-c711-419a-9e24-bdb4c6e9aef6-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-chzp2\" (UID: \"1b632d84-c711-419a-9e24-bdb4c6e9aef6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chzp2" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.396587 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1d9869e2-6f55-4246-8ed0-b8af9dab3f74-bound-sa-token\") pod \"ingress-operator-5b745b69d9-r6cxt\" (UID: \"1d9869e2-6f55-4246-8ed0-b8af9dab3f74\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.396633 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6f4x\" (UniqueName: \"kubernetes.io/projected/7a1dfb55-8680-4cbe-bd78-caca2e847caf-kube-api-access-k6f4x\") pod \"router-default-5444994796-lz6mw\" (UID: \"7a1dfb55-8680-4cbe-bd78-caca2e847caf\") " pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.396675 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/adc3900b-dce0-4da4-bfc2-bca85b2395b2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-pbrmk\" (UID: \"adc3900b-dce0-4da4-bfc2-bca85b2395b2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pbrmk" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.396707 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a1dfb55-8680-4cbe-bd78-caca2e847caf-service-ca-bundle\") pod \"router-default-5444994796-lz6mw\" (UID: \"7a1dfb55-8680-4cbe-bd78-caca2e847caf\") " pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.396732 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/38f843fe-c20b-4bc7-8f45-4ccd4b7be5a5-srv-cert\") pod \"olm-operator-6b444d44fb-ljnh7\" (UID: \"38f843fe-c20b-4bc7-8f45-4ccd4b7be5a5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ljnh7" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.396758 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/714cef39-2960-4a25-ac81-a4e65a115eb3-socket-dir\") pod \"csi-hostpathplugin-4nghl\" (UID: \"714cef39-2960-4a25-ac81-a4e65a115eb3\") " pod="hostpath-provisioner/csi-hostpathplugin-4nghl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.396784 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/714cef39-2960-4a25-ac81-a4e65a115eb3-csi-data-dir\") pod \"csi-hostpathplugin-4nghl\" (UID: \"714cef39-2960-4a25-ac81-a4e65a115eb3\") " pod="hostpath-provisioner/csi-hostpathplugin-4nghl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.396828 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cc84f60e-094e-4924-b6f1-f0a8ab81aa4e-profile-collector-cert\") pod \"catalog-operator-68c6474976-jzg2m\" (UID: \"cc84f60e-094e-4924-b6f1-f0a8ab81aa4e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jzg2m" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.396877 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n67sj\" (UniqueName: \"kubernetes.io/projected/2fa1ede8-3ea3-421d-929d-f6bf9cc1db0e-kube-api-access-n67sj\") pod \"migrator-59844c95c7-xtzbx\" (UID: \"2fa1ede8-3ea3-421d-929d-f6bf9cc1db0e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xtzbx" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.396907 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f18250a8-66c1-445d-9452-081de13b24f7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-zkbcz\" (UID: \"f18250a8-66c1-445d-9452-081de13b24f7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zkbcz" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.396977 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1d9869e2-6f55-4246-8ed0-b8af9dab3f74-metrics-tls\") pod \"ingress-operator-5b745b69d9-r6cxt\" (UID: \"1d9869e2-6f55-4246-8ed0-b8af9dab3f74\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.397003 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fcaee576-dff0-4a67-a0b1-7347b3030729-config-volume\") pod \"dns-default-jcdhl\" (UID: \"fcaee576-dff0-4a67-a0b1-7347b3030729\") " pod="openshift-dns/dns-default-jcdhl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.397183 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/706713ee-0ea2-4018-847c-ccf3a0fafb1c-signing-key\") pod \"service-ca-9c57cc56f-kfqcf\" (UID: \"706713ee-0ea2-4018-847c-ccf3a0fafb1c\") " pod="openshift-service-ca/service-ca-9c57cc56f-kfqcf" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.397228 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a1dfb55-8680-4cbe-bd78-caca2e847caf-metrics-certs\") pod \"router-default-5444994796-lz6mw\" (UID: \"7a1dfb55-8680-4cbe-bd78-caca2e847caf\") " pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.397251 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh8t8\" (UniqueName: \"kubernetes.io/projected/77b31235-8b07-4d66-aec8-64e5b7fae08e-kube-api-access-nh8t8\") pod \"service-ca-operator-777779d784-44gfk\" (UID: \"77b31235-8b07-4d66-aec8-64e5b7fae08e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-44gfk" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.397282 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1d9869e2-6f55-4246-8ed0-b8af9dab3f74-trusted-ca\") pod \"ingress-operator-5b745b69d9-r6cxt\" (UID: \"1d9869e2-6f55-4246-8ed0-b8af9dab3f74\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.397300 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1b632d84-c711-419a-9e24-bdb4c6e9aef6-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-chzp2\" (UID: \"1b632d84-c711-419a-9e24-bdb4c6e9aef6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chzp2" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.397465 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a1dfb55-8680-4cbe-bd78-caca2e847caf-service-ca-bundle\") pod \"router-default-5444994796-lz6mw\" (UID: \"7a1dfb55-8680-4cbe-bd78-caca2e847caf\") " pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.397803 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nzm9\" (UniqueName: \"kubernetes.io/projected/706713ee-0ea2-4018-847c-ccf3a0fafb1c-kube-api-access-8nzm9\") pod \"service-ca-9c57cc56f-kfqcf\" (UID: \"706713ee-0ea2-4018-847c-ccf3a0fafb1c\") " pod="openshift-service-ca/service-ca-9c57cc56f-kfqcf" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.397831 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/80515d06-c09e-4c9d-a90f-43cc84edf4c9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-l2t56\" (UID: \"80515d06-c09e-4c9d-a90f-43cc84edf4c9\") " pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.397850 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/706713ee-0ea2-4018-847c-ccf3a0fafb1c-signing-cabundle\") pod \"service-ca-9c57cc56f-kfqcf\" (UID: \"706713ee-0ea2-4018-847c-ccf3a0fafb1c\") " pod="openshift-service-ca/service-ca-9c57cc56f-kfqcf" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.397967 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1fd5b799-74c2-4ffa-b3d9-6745c66ba28f-certs\") pod \"machine-config-server-zxb4v\" (UID: \"1fd5b799-74c2-4ffa-b3d9-6745c66ba28f\") " pod="openshift-machine-config-operator/machine-config-server-zxb4v" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.397998 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh5km\" (UniqueName: \"kubernetes.io/projected/48be8ad8-4c02-4bea-a143-449763b39d54-kube-api-access-wh5km\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.398019 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77b31235-8b07-4d66-aec8-64e5b7fae08e-config\") pod \"service-ca-operator-777779d784-44gfk\" (UID: \"77b31235-8b07-4d66-aec8-64e5b7fae08e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-44gfk" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.399488 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7a1dfb55-8680-4cbe-bd78-caca2e847caf-stats-auth\") pod \"router-default-5444994796-lz6mw\" (UID: \"7a1dfb55-8680-4cbe-bd78-caca2e847caf\") " pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.399745 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab-apiservice-cert\") pod \"packageserver-d55dfcdfc-v68nn\" (UID: \"2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.399775 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/48be8ad8-4c02-4bea-a143-449763b39d54-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.399839 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/48be8ad8-4c02-4bea-a143-449763b39d54-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.399884 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fcaee576-dff0-4a67-a0b1-7347b3030729-metrics-tls\") pod \"dns-default-jcdhl\" (UID: \"fcaee576-dff0-4a67-a0b1-7347b3030729\") " pod="openshift-dns/dns-default-jcdhl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.399890 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cc84f60e-094e-4924-b6f1-f0a8ab81aa4e-srv-cert\") pod \"catalog-operator-68c6474976-jzg2m\" (UID: \"cc84f60e-094e-4924-b6f1-f0a8ab81aa4e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jzg2m" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.401780 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02158b16-2eb1-4b8e-b1bb-55285b22d053-cert\") pod \"ingress-canary-vhtdt\" (UID: \"02158b16-2eb1-4b8e-b1bb-55285b22d053\") " pod="openshift-ingress-canary/ingress-canary-vhtdt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.401824 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.401906 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cc84f60e-094e-4924-b6f1-f0a8ab81aa4e-profile-collector-cert\") pod \"catalog-operator-68c6474976-jzg2m\" (UID: \"cc84f60e-094e-4924-b6f1-f0a8ab81aa4e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jzg2m" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.402098 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g6v4\" (UniqueName: \"kubernetes.io/projected/adc3900b-dce0-4da4-bfc2-bca85b2395b2-kube-api-access-4g6v4\") pod \"package-server-manager-789f6589d5-pbrmk\" (UID: \"adc3900b-dce0-4da4-bfc2-bca85b2395b2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pbrmk" Jan 29 16:12:16 crc kubenswrapper[4714]: E0129 16:12:16.402197 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:16.902166629 +0000 UTC m=+143.422667749 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.402207 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a1dfb55-8680-4cbe-bd78-caca2e847caf-metrics-certs\") pod \"router-default-5444994796-lz6mw\" (UID: \"7a1dfb55-8680-4cbe-bd78-caca2e847caf\") " pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.402278 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/714cef39-2960-4a25-ac81-a4e65a115eb3-mountpoint-dir\") pod \"csi-hostpathplugin-4nghl\" (UID: \"714cef39-2960-4a25-ac81-a4e65a115eb3\") " pod="hostpath-provisioner/csi-hostpathplugin-4nghl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.402969 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b7cf219f-4e80-47fc-b349-ea5c7eab6d9d-metrics-tls\") pod \"dns-operator-744455d44c-sv7xw\" (UID: \"b7cf219f-4e80-47fc-b349-ea5c7eab6d9d\") " pod="openshift-dns-operator/dns-operator-744455d44c-sv7xw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.403441 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7a1dfb55-8680-4cbe-bd78-caca2e847caf-default-certificate\") pod \"router-default-5444994796-lz6mw\" (UID: \"7a1dfb55-8680-4cbe-bd78-caca2e847caf\") " pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.403507 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/714cef39-2960-4a25-ac81-a4e65a115eb3-registration-dir\") pod \"csi-hostpathplugin-4nghl\" (UID: \"714cef39-2960-4a25-ac81-a4e65a115eb3\") " pod="hostpath-provisioner/csi-hostpathplugin-4nghl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.403541 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77b31235-8b07-4d66-aec8-64e5b7fae08e-serving-cert\") pod \"service-ca-operator-777779d784-44gfk\" (UID: \"77b31235-8b07-4d66-aec8-64e5b7fae08e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-44gfk" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.404085 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfkhr\" (UniqueName: \"kubernetes.io/projected/1d9869e2-6f55-4246-8ed0-b8af9dab3f74-kube-api-access-gfkhr\") pod \"ingress-operator-5b745b69d9-r6cxt\" (UID: \"1d9869e2-6f55-4246-8ed0-b8af9dab3f74\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.404160 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80515d06-c09e-4c9d-a90f-43cc84edf4c9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-l2t56\" (UID: \"80515d06-c09e-4c9d-a90f-43cc84edf4c9\") " pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.404544 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsbnv\" (UniqueName: \"kubernetes.io/projected/02158b16-2eb1-4b8e-b1bb-55285b22d053-kube-api-access-qsbnv\") pod \"ingress-canary-vhtdt\" (UID: \"02158b16-2eb1-4b8e-b1bb-55285b22d053\") " pod="openshift-ingress-canary/ingress-canary-vhtdt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.404598 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfc66\" (UniqueName: \"kubernetes.io/projected/80515d06-c09e-4c9d-a90f-43cc84edf4c9-kube-api-access-xfc66\") pod \"marketplace-operator-79b997595-l2t56\" (UID: \"80515d06-c09e-4c9d-a90f-43cc84edf4c9\") " pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.404663 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab-webhook-cert\") pod \"packageserver-d55dfcdfc-v68nn\" (UID: \"2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.405770 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80515d06-c09e-4c9d-a90f-43cc84edf4c9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-l2t56\" (UID: \"80515d06-c09e-4c9d-a90f-43cc84edf4c9\") " pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.405859 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b7cf219f-4e80-47fc-b349-ea5c7eab6d9d-metrics-tls\") pod \"dns-operator-744455d44c-sv7xw\" (UID: \"b7cf219f-4e80-47fc-b349-ea5c7eab6d9d\") " pod="openshift-dns-operator/dns-operator-744455d44c-sv7xw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.406748 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/80515d06-c09e-4c9d-a90f-43cc84edf4c9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-l2t56\" (UID: \"80515d06-c09e-4c9d-a90f-43cc84edf4c9\") " pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.420698 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfz7c\" (UniqueName: \"kubernetes.io/projected/cc84f60e-094e-4924-b6f1-f0a8ab81aa4e-kube-api-access-zfz7c\") pod \"catalog-operator-68c6474976-jzg2m\" (UID: \"cc84f60e-094e-4924-b6f1-f0a8ab81aa4e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jzg2m" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.429824 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nf7jb" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.442793 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjnvs\" (UniqueName: \"kubernetes.io/projected/b7cf219f-4e80-47fc-b349-ea5c7eab6d9d-kube-api-access-wjnvs\") pod \"dns-operator-744455d44c-sv7xw\" (UID: \"b7cf219f-4e80-47fc-b349-ea5c7eab6d9d\") " pod="openshift-dns-operator/dns-operator-744455d44c-sv7xw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.443511 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw"] Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.454706 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-sv7xw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.454822 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7a1dfb55-8680-4cbe-bd78-caca2e847caf-default-certificate\") pod \"router-default-5444994796-lz6mw\" (UID: \"7a1dfb55-8680-4cbe-bd78-caca2e847caf\") " pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.456071 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/48be8ad8-4c02-4bea-a143-449763b39d54-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.456855 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/48be8ad8-4c02-4bea-a143-449763b39d54-trusted-ca\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.458226 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/48be8ad8-4c02-4bea-a143-449763b39d54-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.458362 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/48be8ad8-4c02-4bea-a143-449763b39d54-registry-tls\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.466997 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1d9869e2-6f55-4246-8ed0-b8af9dab3f74-trusted-ca\") pod \"ingress-operator-5b745b69d9-r6cxt\" (UID: \"1d9869e2-6f55-4246-8ed0-b8af9dab3f74\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.469660 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1d9869e2-6f55-4246-8ed0-b8af9dab3f74-metrics-tls\") pod \"ingress-operator-5b745b69d9-r6cxt\" (UID: \"1d9869e2-6f55-4246-8ed0-b8af9dab3f74\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.470358 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/48be8ad8-4c02-4bea-a143-449763b39d54-bound-sa-token\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: W0129 16:12:16.472352 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f71ba3e_c687_4ff7_9475_1e18ded764f6.slice/crio-91c8d3bea295ee12e93d224413f942892c3f7aeff1c8def01e65177720debd20 WatchSource:0}: Error finding container 91c8d3bea295ee12e93d224413f942892c3f7aeff1c8def01e65177720debd20: Status 404 returned error can't find the container with id 91c8d3bea295ee12e93d224413f942892c3f7aeff1c8def01e65177720debd20 Jan 29 16:12:16 crc kubenswrapper[4714]: W0129 16:12:16.476073 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0387af3d_8796_46b0_9282_9ecbda7fe3a7.slice/crio-aeff327d5990523e2cb090f23196b9fa5cb535b0bfdecedfb353339232ef1474 WatchSource:0}: Error finding container aeff327d5990523e2cb090f23196b9fa5cb535b0bfdecedfb353339232ef1474: Status 404 returned error can't find the container with id aeff327d5990523e2cb090f23196b9fa5cb535b0bfdecedfb353339232ef1474 Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.483060 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1d9869e2-6f55-4246-8ed0-b8af9dab3f74-bound-sa-token\") pod \"ingress-operator-5b745b69d9-r6cxt\" (UID: \"1d9869e2-6f55-4246-8ed0-b8af9dab3f74\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.505444 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:16 crc kubenswrapper[4714]: E0129 16:12:16.505585 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:17.00556165 +0000 UTC m=+143.526062760 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.505711 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f18250a8-66c1-445d-9452-081de13b24f7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-zkbcz\" (UID: \"f18250a8-66c1-445d-9452-081de13b24f7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zkbcz" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.505743 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fcaee576-dff0-4a67-a0b1-7347b3030729-config-volume\") pod \"dns-default-jcdhl\" (UID: \"fcaee576-dff0-4a67-a0b1-7347b3030729\") " pod="openshift-dns/dns-default-jcdhl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.505769 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/706713ee-0ea2-4018-847c-ccf3a0fafb1c-signing-key\") pod \"service-ca-9c57cc56f-kfqcf\" (UID: \"706713ee-0ea2-4018-847c-ccf3a0fafb1c\") " pod="openshift-service-ca/service-ca-9c57cc56f-kfqcf" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.505793 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh8t8\" (UniqueName: \"kubernetes.io/projected/77b31235-8b07-4d66-aec8-64e5b7fae08e-kube-api-access-nh8t8\") pod \"service-ca-operator-777779d784-44gfk\" (UID: \"77b31235-8b07-4d66-aec8-64e5b7fae08e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-44gfk" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.505817 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1b632d84-c711-419a-9e24-bdb4c6e9aef6-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-chzp2\" (UID: \"1b632d84-c711-419a-9e24-bdb4c6e9aef6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chzp2" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.505839 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nzm9\" (UniqueName: \"kubernetes.io/projected/706713ee-0ea2-4018-847c-ccf3a0fafb1c-kube-api-access-8nzm9\") pod \"service-ca-9c57cc56f-kfqcf\" (UID: \"706713ee-0ea2-4018-847c-ccf3a0fafb1c\") " pod="openshift-service-ca/service-ca-9c57cc56f-kfqcf" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.505861 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/706713ee-0ea2-4018-847c-ccf3a0fafb1c-signing-cabundle\") pod \"service-ca-9c57cc56f-kfqcf\" (UID: \"706713ee-0ea2-4018-847c-ccf3a0fafb1c\") " pod="openshift-service-ca/service-ca-9c57cc56f-kfqcf" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.505890 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1fd5b799-74c2-4ffa-b3d9-6745c66ba28f-certs\") pod \"machine-config-server-zxb4v\" (UID: \"1fd5b799-74c2-4ffa-b3d9-6745c66ba28f\") " pod="openshift-machine-config-operator/machine-config-server-zxb4v" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.505912 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77b31235-8b07-4d66-aec8-64e5b7fae08e-config\") pod \"service-ca-operator-777779d784-44gfk\" (UID: \"77b31235-8b07-4d66-aec8-64e5b7fae08e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-44gfk" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.505939 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab-apiservice-cert\") pod \"packageserver-d55dfcdfc-v68nn\" (UID: \"2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.505976 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fcaee576-dff0-4a67-a0b1-7347b3030729-metrics-tls\") pod \"dns-default-jcdhl\" (UID: \"fcaee576-dff0-4a67-a0b1-7347b3030729\") " pod="openshift-dns/dns-default-jcdhl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506009 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02158b16-2eb1-4b8e-b1bb-55285b22d053-cert\") pod \"ingress-canary-vhtdt\" (UID: \"02158b16-2eb1-4b8e-b1bb-55285b22d053\") " pod="openshift-ingress-canary/ingress-canary-vhtdt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506036 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506059 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g6v4\" (UniqueName: \"kubernetes.io/projected/adc3900b-dce0-4da4-bfc2-bca85b2395b2-kube-api-access-4g6v4\") pod \"package-server-manager-789f6589d5-pbrmk\" (UID: \"adc3900b-dce0-4da4-bfc2-bca85b2395b2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pbrmk" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506092 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/714cef39-2960-4a25-ac81-a4e65a115eb3-mountpoint-dir\") pod \"csi-hostpathplugin-4nghl\" (UID: \"714cef39-2960-4a25-ac81-a4e65a115eb3\") " pod="hostpath-provisioner/csi-hostpathplugin-4nghl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506130 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77b31235-8b07-4d66-aec8-64e5b7fae08e-serving-cert\") pod \"service-ca-operator-777779d784-44gfk\" (UID: \"77b31235-8b07-4d66-aec8-64e5b7fae08e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-44gfk" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506152 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/714cef39-2960-4a25-ac81-a4e65a115eb3-registration-dir\") pod \"csi-hostpathplugin-4nghl\" (UID: \"714cef39-2960-4a25-ac81-a4e65a115eb3\") " pod="hostpath-provisioner/csi-hostpathplugin-4nghl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506186 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsbnv\" (UniqueName: \"kubernetes.io/projected/02158b16-2eb1-4b8e-b1bb-55285b22d053-kube-api-access-qsbnv\") pod \"ingress-canary-vhtdt\" (UID: \"02158b16-2eb1-4b8e-b1bb-55285b22d053\") " pod="openshift-ingress-canary/ingress-canary-vhtdt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506221 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab-webhook-cert\") pod \"packageserver-d55dfcdfc-v68nn\" (UID: \"2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506258 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b632d84-c711-419a-9e24-bdb4c6e9aef6-config\") pod \"kube-apiserver-operator-766d6c64bb-chzp2\" (UID: \"1b632d84-c711-419a-9e24-bdb4c6e9aef6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chzp2" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506284 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/38f843fe-c20b-4bc7-8f45-4ccd4b7be5a5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-ljnh7\" (UID: \"38f843fe-c20b-4bc7-8f45-4ccd4b7be5a5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ljnh7" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506308 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45vkr\" (UniqueName: \"kubernetes.io/projected/fcaee576-dff0-4a67-a0b1-7347b3030729-kube-api-access-45vkr\") pod \"dns-default-jcdhl\" (UID: \"fcaee576-dff0-4a67-a0b1-7347b3030729\") " pod="openshift-dns/dns-default-jcdhl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506330 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/714cef39-2960-4a25-ac81-a4e65a115eb3-plugins-dir\") pod \"csi-hostpathplugin-4nghl\" (UID: \"714cef39-2960-4a25-ac81-a4e65a115eb3\") " pod="hostpath-provisioner/csi-hostpathplugin-4nghl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506355 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fzjx\" (UniqueName: \"kubernetes.io/projected/cb979c55-3027-4d92-94b9-cd17c32e6331-kube-api-access-7fzjx\") pod \"collect-profiles-29495040-5mkf8\" (UID: \"cb979c55-3027-4d92-94b9-cd17c32e6331\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506378 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqcjl\" (UniqueName: \"kubernetes.io/projected/38f843fe-c20b-4bc7-8f45-4ccd4b7be5a5-kube-api-access-fqcjl\") pod \"olm-operator-6b444d44fb-ljnh7\" (UID: \"38f843fe-c20b-4bc7-8f45-4ccd4b7be5a5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ljnh7" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506399 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb979c55-3027-4d92-94b9-cd17c32e6331-secret-volume\") pod \"collect-profiles-29495040-5mkf8\" (UID: \"cb979c55-3027-4d92-94b9-cd17c32e6331\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506424 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf5f9\" (UniqueName: \"kubernetes.io/projected/f18250a8-66c1-445d-9452-081de13b24f7-kube-api-access-lf5f9\") pod \"multus-admission-controller-857f4d67dd-zkbcz\" (UID: \"f18250a8-66c1-445d-9452-081de13b24f7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zkbcz" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506448 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfkhq\" (UniqueName: \"kubernetes.io/projected/1fd5b799-74c2-4ffa-b3d9-6745c66ba28f-kube-api-access-vfkhq\") pod \"machine-config-server-zxb4v\" (UID: \"1fd5b799-74c2-4ffa-b3d9-6745c66ba28f\") " pod="openshift-machine-config-operator/machine-config-server-zxb4v" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506478 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6k2j\" (UniqueName: \"kubernetes.io/projected/8062d225-aa57-48df-bf28-2254ecc4f635-kube-api-access-z6k2j\") pod \"control-plane-machine-set-operator-78cbb6b69f-sq9mx\" (UID: \"8062d225-aa57-48df-bf28-2254ecc4f635\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sq9mx" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506500 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab-tmpfs\") pod \"packageserver-d55dfcdfc-v68nn\" (UID: \"2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506523 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1fd5b799-74c2-4ffa-b3d9-6745c66ba28f-node-bootstrap-token\") pod \"machine-config-server-zxb4v\" (UID: \"1fd5b799-74c2-4ffa-b3d9-6745c66ba28f\") " pod="openshift-machine-config-operator/machine-config-server-zxb4v" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506545 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p4cl\" (UniqueName: \"kubernetes.io/projected/2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab-kube-api-access-4p4cl\") pod \"packageserver-d55dfcdfc-v68nn\" (UID: \"2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506569 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8062d225-aa57-48df-bf28-2254ecc4f635-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sq9mx\" (UID: \"8062d225-aa57-48df-bf28-2254ecc4f635\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sq9mx" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506594 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwsfq\" (UniqueName: \"kubernetes.io/projected/714cef39-2960-4a25-ac81-a4e65a115eb3-kube-api-access-xwsfq\") pod \"csi-hostpathplugin-4nghl\" (UID: \"714cef39-2960-4a25-ac81-a4e65a115eb3\") " pod="hostpath-provisioner/csi-hostpathplugin-4nghl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506615 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb979c55-3027-4d92-94b9-cd17c32e6331-config-volume\") pod \"collect-profiles-29495040-5mkf8\" (UID: \"cb979c55-3027-4d92-94b9-cd17c32e6331\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506657 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b632d84-c711-419a-9e24-bdb4c6e9aef6-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-chzp2\" (UID: \"1b632d84-c711-419a-9e24-bdb4c6e9aef6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chzp2" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506688 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/adc3900b-dce0-4da4-bfc2-bca85b2395b2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-pbrmk\" (UID: \"adc3900b-dce0-4da4-bfc2-bca85b2395b2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pbrmk" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506712 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/714cef39-2960-4a25-ac81-a4e65a115eb3-socket-dir\") pod \"csi-hostpathplugin-4nghl\" (UID: \"714cef39-2960-4a25-ac81-a4e65a115eb3\") " pod="hostpath-provisioner/csi-hostpathplugin-4nghl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506736 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/38f843fe-c20b-4bc7-8f45-4ccd4b7be5a5-srv-cert\") pod \"olm-operator-6b444d44fb-ljnh7\" (UID: \"38f843fe-c20b-4bc7-8f45-4ccd4b7be5a5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ljnh7" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506760 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/714cef39-2960-4a25-ac81-a4e65a115eb3-csi-data-dir\") pod \"csi-hostpathplugin-4nghl\" (UID: \"714cef39-2960-4a25-ac81-a4e65a115eb3\") " pod="hostpath-provisioner/csi-hostpathplugin-4nghl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.506801 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fcaee576-dff0-4a67-a0b1-7347b3030729-config-volume\") pod \"dns-default-jcdhl\" (UID: \"fcaee576-dff0-4a67-a0b1-7347b3030729\") " pod="openshift-dns/dns-default-jcdhl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.507191 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/714cef39-2960-4a25-ac81-a4e65a115eb3-plugins-dir\") pod \"csi-hostpathplugin-4nghl\" (UID: \"714cef39-2960-4a25-ac81-a4e65a115eb3\") " pod="hostpath-provisioner/csi-hostpathplugin-4nghl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.507477 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/714cef39-2960-4a25-ac81-a4e65a115eb3-mountpoint-dir\") pod \"csi-hostpathplugin-4nghl\" (UID: \"714cef39-2960-4a25-ac81-a4e65a115eb3\") " pod="hostpath-provisioner/csi-hostpathplugin-4nghl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.507916 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/706713ee-0ea2-4018-847c-ccf3a0fafb1c-signing-cabundle\") pod \"service-ca-9c57cc56f-kfqcf\" (UID: \"706713ee-0ea2-4018-847c-ccf3a0fafb1c\") " pod="openshift-service-ca/service-ca-9c57cc56f-kfqcf" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.508527 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/714cef39-2960-4a25-ac81-a4e65a115eb3-socket-dir\") pod \"csi-hostpathplugin-4nghl\" (UID: \"714cef39-2960-4a25-ac81-a4e65a115eb3\") " pod="hostpath-provisioner/csi-hostpathplugin-4nghl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.508687 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb979c55-3027-4d92-94b9-cd17c32e6331-config-volume\") pod \"collect-profiles-29495040-5mkf8\" (UID: \"cb979c55-3027-4d92-94b9-cd17c32e6331\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.510049 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77b31235-8b07-4d66-aec8-64e5b7fae08e-config\") pod \"service-ca-operator-777779d784-44gfk\" (UID: \"77b31235-8b07-4d66-aec8-64e5b7fae08e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-44gfk" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.510453 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/714cef39-2960-4a25-ac81-a4e65a115eb3-registration-dir\") pod \"csi-hostpathplugin-4nghl\" (UID: \"714cef39-2960-4a25-ac81-a4e65a115eb3\") " pod="hostpath-provisioner/csi-hostpathplugin-4nghl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.510973 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab-tmpfs\") pod \"packageserver-d55dfcdfc-v68nn\" (UID: \"2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" Jan 29 16:12:16 crc kubenswrapper[4714]: E0129 16:12:16.510925 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:17.010906344 +0000 UTC m=+143.531407464 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.511106 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/714cef39-2960-4a25-ac81-a4e65a115eb3-csi-data-dir\") pod \"csi-hostpathplugin-4nghl\" (UID: \"714cef39-2960-4a25-ac81-a4e65a115eb3\") " pod="hostpath-provisioner/csi-hostpathplugin-4nghl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.511268 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/706713ee-0ea2-4018-847c-ccf3a0fafb1c-signing-key\") pod \"service-ca-9c57cc56f-kfqcf\" (UID: \"706713ee-0ea2-4018-847c-ccf3a0fafb1c\") " pod="openshift-service-ca/service-ca-9c57cc56f-kfqcf" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.511292 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b632d84-c711-419a-9e24-bdb4c6e9aef6-config\") pod \"kube-apiserver-operator-766d6c64bb-chzp2\" (UID: \"1b632d84-c711-419a-9e24-bdb4c6e9aef6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chzp2" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.511539 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fcaee576-dff0-4a67-a0b1-7347b3030729-metrics-tls\") pod \"dns-default-jcdhl\" (UID: \"fcaee576-dff0-4a67-a0b1-7347b3030729\") " pod="openshift-dns/dns-default-jcdhl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.511471 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b632d84-c711-419a-9e24-bdb4c6e9aef6-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-chzp2\" (UID: \"1b632d84-c711-419a-9e24-bdb4c6e9aef6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chzp2" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.511711 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f18250a8-66c1-445d-9452-081de13b24f7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-zkbcz\" (UID: \"f18250a8-66c1-445d-9452-081de13b24f7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zkbcz" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.512069 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jzg2m" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.512256 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77b31235-8b07-4d66-aec8-64e5b7fae08e-serving-cert\") pod \"service-ca-operator-777779d784-44gfk\" (UID: \"77b31235-8b07-4d66-aec8-64e5b7fae08e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-44gfk" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.513174 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/38f843fe-c20b-4bc7-8f45-4ccd4b7be5a5-srv-cert\") pod \"olm-operator-6b444d44fb-ljnh7\" (UID: \"38f843fe-c20b-4bc7-8f45-4ccd4b7be5a5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ljnh7" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.513423 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8062d225-aa57-48df-bf28-2254ecc4f635-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sq9mx\" (UID: \"8062d225-aa57-48df-bf28-2254ecc4f635\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sq9mx" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.513617 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1fd5b799-74c2-4ffa-b3d9-6745c66ba28f-certs\") pod \"machine-config-server-zxb4v\" (UID: \"1fd5b799-74c2-4ffa-b3d9-6745c66ba28f\") " pod="openshift-machine-config-operator/machine-config-server-zxb4v" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.514183 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6f4x\" (UniqueName: \"kubernetes.io/projected/7a1dfb55-8680-4cbe-bd78-caca2e847caf-kube-api-access-k6f4x\") pod \"router-default-5444994796-lz6mw\" (UID: \"7a1dfb55-8680-4cbe-bd78-caca2e847caf\") " pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.514385 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1fd5b799-74c2-4ffa-b3d9-6745c66ba28f-node-bootstrap-token\") pod \"machine-config-server-zxb4v\" (UID: \"1fd5b799-74c2-4ffa-b3d9-6745c66ba28f\") " pod="openshift-machine-config-operator/machine-config-server-zxb4v" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.514593 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02158b16-2eb1-4b8e-b1bb-55285b22d053-cert\") pod \"ingress-canary-vhtdt\" (UID: \"02158b16-2eb1-4b8e-b1bb-55285b22d053\") " pod="openshift-ingress-canary/ingress-canary-vhtdt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.514691 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/38f843fe-c20b-4bc7-8f45-4ccd4b7be5a5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-ljnh7\" (UID: \"38f843fe-c20b-4bc7-8f45-4ccd4b7be5a5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ljnh7" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.516250 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab-webhook-cert\") pod \"packageserver-d55dfcdfc-v68nn\" (UID: \"2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.516875 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/adc3900b-dce0-4da4-bfc2-bca85b2395b2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-pbrmk\" (UID: \"adc3900b-dce0-4da4-bfc2-bca85b2395b2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pbrmk" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.517596 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab-apiservice-cert\") pod \"packageserver-d55dfcdfc-v68nn\" (UID: \"2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.544927 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n67sj\" (UniqueName: \"kubernetes.io/projected/2fa1ede8-3ea3-421d-929d-f6bf9cc1db0e-kube-api-access-n67sj\") pod \"migrator-59844c95c7-xtzbx\" (UID: \"2fa1ede8-3ea3-421d-929d-f6bf9cc1db0e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xtzbx" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.577926 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh5km\" (UniqueName: \"kubernetes.io/projected/48be8ad8-4c02-4bea-a143-449763b39d54-kube-api-access-wh5km\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.586675 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfkhr\" (UniqueName: \"kubernetes.io/projected/1d9869e2-6f55-4246-8ed0-b8af9dab3f74-kube-api-access-gfkhr\") pod \"ingress-operator-5b745b69d9-r6cxt\" (UID: \"1d9869e2-6f55-4246-8ed0-b8af9dab3f74\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.607555 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:16 crc kubenswrapper[4714]: E0129 16:12:16.607769 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:17.107752475 +0000 UTC m=+143.628253585 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.608120 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfc66\" (UniqueName: \"kubernetes.io/projected/80515d06-c09e-4c9d-a90f-43cc84edf4c9-kube-api-access-xfc66\") pod \"marketplace-operator-79b997595-l2t56\" (UID: \"80515d06-c09e-4c9d-a90f-43cc84edf4c9\") " pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.608234 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: E0129 16:12:16.609175 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:17.108637612 +0000 UTC m=+143.629138732 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.624694 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6jl75"] Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.628200 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-99knh"] Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.643913 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1b632d84-c711-419a-9e24-bdb4c6e9aef6-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-chzp2\" (UID: \"1b632d84-c711-419a-9e24-bdb4c6e9aef6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chzp2" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.665273 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fzjx\" (UniqueName: \"kubernetes.io/projected/cb979c55-3027-4d92-94b9-cd17c32e6331-kube-api-access-7fzjx\") pod \"collect-profiles-29495040-5mkf8\" (UID: \"cb979c55-3027-4d92-94b9-cd17c32e6331\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.669766 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-z4h55"] Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.682468 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqcjl\" (UniqueName: \"kubernetes.io/projected/38f843fe-c20b-4bc7-8f45-4ccd4b7be5a5-kube-api-access-fqcjl\") pod \"olm-operator-6b444d44fb-ljnh7\" (UID: \"38f843fe-c20b-4bc7-8f45-4ccd4b7be5a5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ljnh7" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.701009 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4dn69"] Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.709068 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:16 crc kubenswrapper[4714]: E0129 16:12:16.709177 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:17.209155996 +0000 UTC m=+143.729657116 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.709463 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: E0129 16:12:16.709982 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:17.209973881 +0000 UTC m=+143.730475001 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.728213 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nzm9\" (UniqueName: \"kubernetes.io/projected/706713ee-0ea2-4018-847c-ccf3a0fafb1c-kube-api-access-8nzm9\") pod \"service-ca-9c57cc56f-kfqcf\" (UID: \"706713ee-0ea2-4018-847c-ccf3a0fafb1c\") " pod="openshift-service-ca/service-ca-9c57cc56f-kfqcf" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.736992 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.745651 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwsfq\" (UniqueName: \"kubernetes.io/projected/714cef39-2960-4a25-ac81-a4e65a115eb3-kube-api-access-xwsfq\") pod \"csi-hostpathplugin-4nghl\" (UID: \"714cef39-2960-4a25-ac81-a4e65a115eb3\") " pod="hostpath-provisioner/csi-hostpathplugin-4nghl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.762231 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.763905 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh8t8\" (UniqueName: \"kubernetes.io/projected/77b31235-8b07-4d66-aec8-64e5b7fae08e-kube-api-access-nh8t8\") pod \"service-ca-operator-777779d784-44gfk\" (UID: \"77b31235-8b07-4d66-aec8-64e5b7fae08e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-44gfk" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.768456 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6k2j\" (UniqueName: \"kubernetes.io/projected/8062d225-aa57-48df-bf28-2254ecc4f635-kube-api-access-z6k2j\") pod \"control-plane-machine-set-operator-78cbb6b69f-sq9mx\" (UID: \"8062d225-aa57-48df-bf28-2254ecc4f635\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sq9mx" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.782579 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.785146 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf5f9\" (UniqueName: \"kubernetes.io/projected/f18250a8-66c1-445d-9452-081de13b24f7-kube-api-access-lf5f9\") pod \"multus-admission-controller-857f4d67dd-zkbcz\" (UID: \"f18250a8-66c1-445d-9452-081de13b24f7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zkbcz" Jan 29 16:12:16 crc kubenswrapper[4714]: W0129 16:12:16.789538 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbfb0bd22_cbd8_4ce8_a4f6_86a16dcdeb92.slice/crio-157fd5666164f797770cad3af5809cb593dc5c7c6941d71c40b7576dc2323486 WatchSource:0}: Error finding container 157fd5666164f797770cad3af5809cb593dc5c7c6941d71c40b7576dc2323486: Status 404 returned error can't find the container with id 157fd5666164f797770cad3af5809cb593dc5c7c6941d71c40b7576dc2323486 Jan 29 16:12:16 crc kubenswrapper[4714]: W0129 16:12:16.791929 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6de35940_bef4_4dfa_9a83_08ba29d73399.slice/crio-bab710f268cee7a17d81afe5168b5be33ec92cbaa20fae67c86001bda19ec16b WatchSource:0}: Error finding container bab710f268cee7a17d81afe5168b5be33ec92cbaa20fae67c86001bda19ec16b: Status 404 returned error can't find the container with id bab710f268cee7a17d81afe5168b5be33ec92cbaa20fae67c86001bda19ec16b Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.792117 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xtzbx" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.810252 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:16 crc kubenswrapper[4714]: E0129 16:12:16.810854 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:17.310818275 +0000 UTC m=+143.831319415 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.812610 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/48be8ad8-4c02-4bea-a143-449763b39d54-registry-certificates\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.812790 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chzp2" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.814562 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfkhq\" (UniqueName: \"kubernetes.io/projected/1fd5b799-74c2-4ffa-b3d9-6745c66ba28f-kube-api-access-vfkhq\") pod \"machine-config-server-zxb4v\" (UID: \"1fd5b799-74c2-4ffa-b3d9-6745c66ba28f\") " pod="openshift-machine-config-operator/machine-config-server-zxb4v" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.821776 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb979c55-3027-4d92-94b9-cd17c32e6331-secret-volume\") pod \"collect-profiles-29495040-5mkf8\" (UID: \"cb979c55-3027-4d92-94b9-cd17c32e6331\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.822043 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sq9mx" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.826804 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g6v4\" (UniqueName: \"kubernetes.io/projected/adc3900b-dce0-4da4-bfc2-bca85b2395b2-kube-api-access-4g6v4\") pod \"package-server-manager-789f6589d5-pbrmk\" (UID: \"adc3900b-dce0-4da4-bfc2-bca85b2395b2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pbrmk" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.829051 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-44gfk" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.837322 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-zkbcz" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.852269 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p4cl\" (UniqueName: \"kubernetes.io/projected/2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab-kube-api-access-4p4cl\") pod \"packageserver-d55dfcdfc-v68nn\" (UID: \"2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.852979 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.865523 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pbrmk" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.880802 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ljnh7" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.894167 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsbnv\" (UniqueName: \"kubernetes.io/projected/02158b16-2eb1-4b8e-b1bb-55285b22d053-kube-api-access-qsbnv\") pod \"ingress-canary-vhtdt\" (UID: \"02158b16-2eb1-4b8e-b1bb-55285b22d053\") " pod="openshift-ingress-canary/ingress-canary-vhtdt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.895737 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-kfqcf" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.903414 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-vhtdt" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.913559 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:16 crc kubenswrapper[4714]: E0129 16:12:16.914020 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:17.414003379 +0000 UTC m=+143.934504499 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.921539 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45vkr\" (UniqueName: \"kubernetes.io/projected/fcaee576-dff0-4a67-a0b1-7347b3030729-kube-api-access-45vkr\") pod \"dns-default-jcdhl\" (UID: \"fcaee576-dff0-4a67-a0b1-7347b3030729\") " pod="openshift-dns/dns-default-jcdhl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.921881 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-4nghl" Jan 29 16:12:16 crc kubenswrapper[4714]: I0129 16:12:16.929284 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-zxb4v" Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.008007 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" event={"ID":"3c2d0611-58f8-4a7e-8280-361c80d62802","Type":"ContainerStarted","Data":"997bb46f3e8548114daabdb0676e47c164f03b6651e1e3ef03b31f66106dbebd"} Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.014864 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:17 crc kubenswrapper[4714]: E0129 16:12:17.015339 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:17.515320298 +0000 UTC m=+144.035821418 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.016023 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" event={"ID":"c779f8ba-7614-49f1-be6d-a9e316ec59ba","Type":"ContainerStarted","Data":"3152c7a0341e5d50a15d3a5dff1e3fe3b0fb3f928bd969fbb893bfd4b05c9599"} Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.016899 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-z4h55" event={"ID":"bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92","Type":"ContainerStarted","Data":"157fd5666164f797770cad3af5809cb593dc5c7c6941d71c40b7576dc2323486"} Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.017826 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" event={"ID":"832097a5-4691-42b6-99cc-38679071d5ee","Type":"ContainerStarted","Data":"3a5ee9422c0e8f2bda4f13b1ec7a93ce78a161df42fc1dddfe6f8337aed30775"} Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.025444 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-jb6jw" event={"ID":"b25d77ec-57de-4c2a-b534-e98bf149b92a","Type":"ContainerStarted","Data":"f4d02887cf93fa7440522b162f0c7f7416034125a8251d52f88c2abe6f4c8b5d"} Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.031701 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-kvp9d" event={"ID":"ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec","Type":"ContainerStarted","Data":"fac742e2689e1006f8d65c8e994ad3939bd198f2f9224e4d331d2df4d5693fc3"} Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.047610 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" event={"ID":"8f71ba3e-c687-4ff7-9475-1e18ded764f6","Type":"ContainerStarted","Data":"91c8d3bea295ee12e93d224413f942892c3f7aeff1c8def01e65177720debd20"} Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.049226 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mrprd"] Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.050222 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vcj84" event={"ID":"eacb9f84-018a-4f64-b211-c9bedce50b9e","Type":"ContainerStarted","Data":"ace923f68ad29b1636a1fe40bfc3ee570490853bcb4fd62025c25c9b78f49d58"} Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.052671 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6jl75" event={"ID":"99bab267-639b-48b1-abc4-8c0373200a39","Type":"ContainerStarted","Data":"0e25e4179fbb1ee7f25b933d8c5c0910ffc2e04351daf128945e263f520c2e59"} Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.055139 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp5md" event={"ID":"0387af3d-8796-46b0-9282-9ecbda7fe3a7","Type":"ContainerStarted","Data":"aeff327d5990523e2cb090f23196b9fa5cb535b0bfdecedfb353339232ef1474"} Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.056683 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-fn75b" event={"ID":"42b66dc3-a385-4350-a943-50f062da35f7","Type":"ContainerStarted","Data":"33cc048aa8b0ae0f2921ff20c63d84f7536484b13e9c5a3fee51211e6060d16e"} Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.059676 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" event={"ID":"97cfbecd-36ef-409b-94e9-f607a1fa2c42","Type":"ContainerStarted","Data":"c4aa9f52cc02c48d04509bb33a37850911598242997e7951d059e039ac5a5e8a"} Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.068203 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" event={"ID":"fbfdd647-1d64-4d35-9af2-6dee52b4c860","Type":"ContainerStarted","Data":"79940598fef6f2445dc05d94ab28a7d984953a342201b3331c2b27e4796135a0"} Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.071346 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m2g9h" event={"ID":"0e2a789d-6a90-4d60-881e-9562cd92e0a7","Type":"ContainerStarted","Data":"a924f3005f8d63da614824356d28979437300d10da79ec4af9d993bb04ed4e85"} Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.079558 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4dn69" event={"ID":"6de35940-bef4-4dfa-9a83-08ba29d73399","Type":"ContainerStarted","Data":"bab710f268cee7a17d81afe5168b5be33ec92cbaa20fae67c86001bda19ec16b"} Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.079966 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-sv7xw"] Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.116615 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:17 crc kubenswrapper[4714]: E0129 16:12:17.116984 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:17.616970146 +0000 UTC m=+144.137471266 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.146427 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.157328 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-nf7jb"] Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.161278 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ch6wr"] Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.183853 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-jcdhl" Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.195873 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj"] Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.218440 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:17 crc kubenswrapper[4714]: E0129 16:12:17.218664 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:17.718629074 +0000 UTC m=+144.239130194 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.218862 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:17 crc kubenswrapper[4714]: E0129 16:12:17.219953 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:17.719934024 +0000 UTC m=+144.240435144 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.229514 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2lpzx"] Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.319736 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:17 crc kubenswrapper[4714]: E0129 16:12:17.319879 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:17.819851399 +0000 UTC m=+144.340352519 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.320018 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:17 crc kubenswrapper[4714]: E0129 16:12:17.320426 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:17.820405626 +0000 UTC m=+144.340906766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.421571 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:17 crc kubenswrapper[4714]: E0129 16:12:17.421839 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:17.921770416 +0000 UTC m=+144.442271576 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.422318 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:17 crc kubenswrapper[4714]: E0129 16:12:17.422626 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:17.922615012 +0000 UTC m=+144.443116132 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.523716 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:17 crc kubenswrapper[4714]: E0129 16:12:17.524179 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:18.024127706 +0000 UTC m=+144.544628826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.627924 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:17 crc kubenswrapper[4714]: E0129 16:12:17.628347 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:18.128333132 +0000 UTC m=+144.648834252 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.729531 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:17 crc kubenswrapper[4714]: E0129 16:12:17.732247 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:18.232206478 +0000 UTC m=+144.752707598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.833521 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:17 crc kubenswrapper[4714]: E0129 16:12:17.833850 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:18.333834116 +0000 UTC m=+144.854335236 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.934999 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:17 crc kubenswrapper[4714]: E0129 16:12:17.935105 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:18.435090642 +0000 UTC m=+144.955591752 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.935242 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:17 crc kubenswrapper[4714]: E0129 16:12:17.935590 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:18.435582737 +0000 UTC m=+144.956083857 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:17 crc kubenswrapper[4714]: I0129 16:12:17.981624 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt"] Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.035953 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:18 crc kubenswrapper[4714]: E0129 16:12:18.036119 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:18.53610192 +0000 UTC m=+145.056603040 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.036239 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:18 crc kubenswrapper[4714]: E0129 16:12:18.036500 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:18.536491132 +0000 UTC m=+145.056992252 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.090063 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mrprd" event={"ID":"554abf87-b1ba-45b1-8130-95b40da3b8bf","Type":"ContainerStarted","Data":"20c7457cb1680b638afb8dc71ad2c9731bfe83fe5accc7acd040c18b0e6bb417"} Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.093179 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ch6wr" event={"ID":"fc8e2d06-1cc2-4ea7-8d87-340d28740e20","Type":"ContainerStarted","Data":"80e18daa72619ece41b19b8e484cfe2d06f7ee1c093574382461cc44012bf2d7"} Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.094511 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-lz6mw" event={"ID":"7a1dfb55-8680-4cbe-bd78-caca2e847caf","Type":"ContainerStarted","Data":"40b83831be51662691c2389228703b2e44d2af6c53cf436bd27964284ccf620f"} Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.095583 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xvrxj" event={"ID":"f3e2f962-69e3-4008-a45f-5c35677f7f36","Type":"ContainerStarted","Data":"a78dd6d9c288aa19f7b8882e1352e2d9efbb00bd10a56122908cef032cacc644"} Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.118133 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nf7jb" event={"ID":"2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96","Type":"ContainerStarted","Data":"f0bde486e4e1af6700632172e82ece5f5d8415f67c21552753efbbf51e1770c6"} Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.128281 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" event={"ID":"3c2d0611-58f8-4a7e-8280-361c80d62802","Type":"ContainerStarted","Data":"1a48e7995b6652b74e320bf25807bcc51cd5f49496615ac9028d0bf40f37019e"} Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.129245 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-zxb4v" event={"ID":"1fd5b799-74c2-4ffa-b3d9-6745c66ba28f","Type":"ContainerStarted","Data":"fc27b02212f07458ccb7ea63701f0795f0a6e29b461d101747658b13ad7ff742"} Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.130161 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj" event={"ID":"d288ee23-1753-48f2-ab82-736defe5fe18","Type":"ContainerStarted","Data":"a6a9144e8ecffc012cb3413e6b06c6e60f1f002771aabaefd491cbe1ccb40491"} Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.130704 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-sv7xw" event={"ID":"b7cf219f-4e80-47fc-b349-ea5c7eab6d9d","Type":"ContainerStarted","Data":"8a0e03bbf7225c254f88d01b6109194d7f1c054597dc2a039f2c879dff5e1628"} Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.131532 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-kvp9d" event={"ID":"ad3c7510-ccc3-453a-91ae-b1f2cf88d2ec","Type":"ContainerStarted","Data":"68e3ecada4f8793d5f95b46b812675b515c4359fd9ffb700b0d202e241aedb17"} Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.134596 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-99knh" event={"ID":"5b5da98c-0704-41c7-8563-707f7af93f41","Type":"ContainerStarted","Data":"ca070377a72c571495332a6a2fd95697bd5a37831362a5be412d22918536e812"} Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.136130 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" event={"ID":"c779f8ba-7614-49f1-be6d-a9e316ec59ba","Type":"ContainerStarted","Data":"e783a0f04dc36f9b00a69d43ae0355c9b664ad354b4cdc74b2a094e5b287a951"} Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.136690 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:18 crc kubenswrapper[4714]: E0129 16:12:18.136909 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:18.636888102 +0000 UTC m=+145.157389222 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.137068 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:18 crc kubenswrapper[4714]: E0129 16:12:18.137594 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:18.637579543 +0000 UTC m=+145.158080663 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.137810 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m2g9h" event={"ID":"0e2a789d-6a90-4d60-881e-9562cd92e0a7","Type":"ContainerStarted","Data":"3f377532bfc03c12b7dca263550589c81deb919323d24c655c2ca133fb41dca9"} Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.138467 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2lpzx" event={"ID":"c14fb55e-a42b-46c9-9521-6e8b60235166","Type":"ContainerStarted","Data":"a8ceb66d30404f617cd622aafe231f15ba56287a7cd277803365b749533792a7"} Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.140206 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-jb6jw" event={"ID":"b25d77ec-57de-4c2a-b534-e98bf149b92a","Type":"ContainerStarted","Data":"867f9250072589a8bc3ab77287b756dc175f3a75811f69552fc6d9db0e0a499a"} Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.238794 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:18 crc kubenswrapper[4714]: E0129 16:12:18.239050 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:18.739027625 +0000 UTC m=+145.259528755 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.239146 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:18 crc kubenswrapper[4714]: E0129 16:12:18.239560 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:18.739549191 +0000 UTC m=+145.260050311 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.340090 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:18 crc kubenswrapper[4714]: E0129 16:12:18.340832 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:18.840816018 +0000 UTC m=+145.361317138 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.441694 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:18 crc kubenswrapper[4714]: E0129 16:12:18.442095 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:18.942083144 +0000 UTC m=+145.462584264 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.526253 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jzg2m"] Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.543430 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:18 crc kubenswrapper[4714]: E0129 16:12:18.543762 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:19.043746673 +0000 UTC m=+145.564247793 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.557784 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-l2t56"] Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.562332 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xtzbx"] Jan 29 16:12:18 crc kubenswrapper[4714]: W0129 16:12:18.580348 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc84f60e_094e_4924_b6f1_f0a8ab81aa4e.slice/crio-11084ed18f3b11987189819d9bb3dd492afdb82473bb2a69f517fe602c5a24f7 WatchSource:0}: Error finding container 11084ed18f3b11987189819d9bb3dd492afdb82473bb2a69f517fe602c5a24f7: Status 404 returned error can't find the container with id 11084ed18f3b11987189819d9bb3dd492afdb82473bb2a69f517fe602c5a24f7 Jan 29 16:12:18 crc kubenswrapper[4714]: W0129 16:12:18.619494 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80515d06_c09e_4c9d_a90f_43cc84edf4c9.slice/crio-5880f1855bae3fd6f603655d40b770623f038db9a3cb9db3918877f801567acc WatchSource:0}: Error finding container 5880f1855bae3fd6f603655d40b770623f038db9a3cb9db3918877f801567acc: Status 404 returned error can't find the container with id 5880f1855bae3fd6f603655d40b770623f038db9a3cb9db3918877f801567acc Jan 29 16:12:18 crc kubenswrapper[4714]: W0129 16:12:18.627695 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fa1ede8_3ea3_421d_929d_f6bf9cc1db0e.slice/crio-f4ea3f04681522c185b8823272251a5951ed23cf1e074da3e02d76c59ea7250d WatchSource:0}: Error finding container f4ea3f04681522c185b8823272251a5951ed23cf1e074da3e02d76c59ea7250d: Status 404 returned error can't find the container with id f4ea3f04681522c185b8823272251a5951ed23cf1e074da3e02d76c59ea7250d Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.645280 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:18 crc kubenswrapper[4714]: E0129 16:12:18.647113 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:19.147097993 +0000 UTC m=+145.667599113 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.663611 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-44gfk"] Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.668056 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chzp2"] Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.672442 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ljnh7"] Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.691625 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sq9mx"] Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.695793 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-vhtdt"] Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.702154 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-4nghl"] Jan 29 16:12:18 crc kubenswrapper[4714]: W0129 16:12:18.707000 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77b31235_8b07_4d66_aec8_64e5b7fae08e.slice/crio-b57b92cf1409311fba74582f0e99821728e24bd5d89e8468fe429886e3a64c84 WatchSource:0}: Error finding container b57b92cf1409311fba74582f0e99821728e24bd5d89e8468fe429886e3a64c84: Status 404 returned error can't find the container with id b57b92cf1409311fba74582f0e99821728e24bd5d89e8468fe429886e3a64c84 Jan 29 16:12:18 crc kubenswrapper[4714]: W0129 16:12:18.707779 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b632d84_c711_419a_9e24_bdb4c6e9aef6.slice/crio-2661bab2f985c589314989ca54a7b8417fe38b5661dd2093e413b43b2402ce34 WatchSource:0}: Error finding container 2661bab2f985c589314989ca54a7b8417fe38b5661dd2093e413b43b2402ce34: Status 404 returned error can't find the container with id 2661bab2f985c589314989ca54a7b8417fe38b5661dd2093e413b43b2402ce34 Jan 29 16:12:18 crc kubenswrapper[4714]: W0129 16:12:18.709289 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38f843fe_c20b_4bc7_8f45_4ccd4b7be5a5.slice/crio-92cdec071830256f624ca0c36fdd787e417207e233c1063a8ab14f07ebc95d1f WatchSource:0}: Error finding container 92cdec071830256f624ca0c36fdd787e417207e233c1063a8ab14f07ebc95d1f: Status 404 returned error can't find the container with id 92cdec071830256f624ca0c36fdd787e417207e233c1063a8ab14f07ebc95d1f Jan 29 16:12:18 crc kubenswrapper[4714]: W0129 16:12:18.714110 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8062d225_aa57_48df_bf28_2254ecc4f635.slice/crio-951697dbf7c719b339005199178b35e78824e9709406b5beed446e6875f95f52 WatchSource:0}: Error finding container 951697dbf7c719b339005199178b35e78824e9709406b5beed446e6875f95f52: Status 404 returned error can't find the container with id 951697dbf7c719b339005199178b35e78824e9709406b5beed446e6875f95f52 Jan 29 16:12:18 crc kubenswrapper[4714]: W0129 16:12:18.716391 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02158b16_2eb1_4b8e_b1bb_55285b22d053.slice/crio-33677f709ea45fe339c09a2748d7e700229caf6e46e524c0ee7292bfc9a240d9 WatchSource:0}: Error finding container 33677f709ea45fe339c09a2748d7e700229caf6e46e524c0ee7292bfc9a240d9: Status 404 returned error can't find the container with id 33677f709ea45fe339c09a2748d7e700229caf6e46e524c0ee7292bfc9a240d9 Jan 29 16:12:18 crc kubenswrapper[4714]: W0129 16:12:18.719648 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod714cef39_2960_4a25_ac81_a4e65a115eb3.slice/crio-ea583be98e43e32e2b9eaab9f7b14cfcc6c03d18a9e1e7bc8ee858d8764bccfb WatchSource:0}: Error finding container ea583be98e43e32e2b9eaab9f7b14cfcc6c03d18a9e1e7bc8ee858d8764bccfb: Status 404 returned error can't find the container with id ea583be98e43e32e2b9eaab9f7b14cfcc6c03d18a9e1e7bc8ee858d8764bccfb Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.748822 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:18 crc kubenswrapper[4714]: E0129 16:12:18.748983 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:19.248965068 +0000 UTC m=+145.769466188 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.749059 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:18 crc kubenswrapper[4714]: E0129 16:12:18.749321 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:19.249313838 +0000 UTC m=+145.769814958 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.805157 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-kfqcf"] Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.807522 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8"] Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.814220 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pbrmk"] Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.834202 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-zkbcz"] Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.850175 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:18 crc kubenswrapper[4714]: E0129 16:12:18.850370 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:19.350344057 +0000 UTC m=+145.870845177 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.850622 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:18 crc kubenswrapper[4714]: E0129 16:12:18.850963 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:19.350929035 +0000 UTC m=+145.871430155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.854582 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn"] Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.858753 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-jcdhl"] Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.952032 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:18 crc kubenswrapper[4714]: E0129 16:12:18.952235 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:19.452206672 +0000 UTC m=+145.972707802 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:18 crc kubenswrapper[4714]: I0129 16:12:18.952517 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:18 crc kubenswrapper[4714]: E0129 16:12:18.952892 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:19.452879963 +0000 UTC m=+145.973381083 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:19 crc kubenswrapper[4714]: W0129 16:12:19.054213 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb979c55_3027_4d92_94b9_cd17c32e6331.slice/crio-64dd0aa019d2ff78b021ae7336dcb6b797f2fc222a74d3ec8b1b6722bf60513e WatchSource:0}: Error finding container 64dd0aa019d2ff78b021ae7336dcb6b797f2fc222a74d3ec8b1b6722bf60513e: Status 404 returned error can't find the container with id 64dd0aa019d2ff78b021ae7336dcb6b797f2fc222a74d3ec8b1b6722bf60513e Jan 29 16:12:19 crc kubenswrapper[4714]: W0129 16:12:19.055455 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod706713ee_0ea2_4018_847c_ccf3a0fafb1c.slice/crio-ac69bf925281e62c711bc61e1e7af18c65e142666843c7ebd8119153dbb6c812 WatchSource:0}: Error finding container ac69bf925281e62c711bc61e1e7af18c65e142666843c7ebd8119153dbb6c812: Status 404 returned error can't find the container with id ac69bf925281e62c711bc61e1e7af18c65e142666843c7ebd8119153dbb6c812 Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.055459 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:19 crc kubenswrapper[4714]: E0129 16:12:19.055623 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:19.555591473 +0000 UTC m=+146.076092633 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.056382 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:19 crc kubenswrapper[4714]: E0129 16:12:19.059593 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:19.559568845 +0000 UTC m=+146.080069995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:19 crc kubenswrapper[4714]: W0129 16:12:19.063166 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fbdedcd_e0ce_4fe9_b19a_53d422b1fcab.slice/crio-bd5ddda0b956be079adf5a95904bc40a5bfab574b272bcc0aace47eca5063bd1 WatchSource:0}: Error finding container bd5ddda0b956be079adf5a95904bc40a5bfab574b272bcc0aace47eca5063bd1: Status 404 returned error can't find the container with id bd5ddda0b956be079adf5a95904bc40a5bfab574b272bcc0aace47eca5063bd1 Jan 29 16:12:19 crc kubenswrapper[4714]: W0129 16:12:19.076114 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podadc3900b_dce0_4da4_bfc2_bca85b2395b2.slice/crio-f8050531b7ff6cac579652a7da90a372a37a8c79cc21280f1c40da4fd018f9e9 WatchSource:0}: Error finding container f8050531b7ff6cac579652a7da90a372a37a8c79cc21280f1c40da4fd018f9e9: Status 404 returned error can't find the container with id f8050531b7ff6cac579652a7da90a372a37a8c79cc21280f1c40da4fd018f9e9 Jan 29 16:12:19 crc kubenswrapper[4714]: W0129 16:12:19.079100 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfcaee576_dff0_4a67_a0b1_7347b3030729.slice/crio-03a0c2abcdf4f7b3d6c47dcfce1a4e19bf09e285b4ddaa4cb1ab60d493e59e52 WatchSource:0}: Error finding container 03a0c2abcdf4f7b3d6c47dcfce1a4e19bf09e285b4ddaa4cb1ab60d493e59e52: Status 404 returned error can't find the container with id 03a0c2abcdf4f7b3d6c47dcfce1a4e19bf09e285b4ddaa4cb1ab60d493e59e52 Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.157679 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:19 crc kubenswrapper[4714]: E0129 16:12:19.157906 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:19.65787336 +0000 UTC m=+146.178374520 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.158058 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:19 crc kubenswrapper[4714]: E0129 16:12:19.158375 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:19.658364895 +0000 UTC m=+146.178866015 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.164291 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pbrmk" event={"ID":"adc3900b-dce0-4da4-bfc2-bca85b2395b2","Type":"ContainerStarted","Data":"f8050531b7ff6cac579652a7da90a372a37a8c79cc21280f1c40da4fd018f9e9"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.168010 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" event={"ID":"fbfdd647-1d64-4d35-9af2-6dee52b4c860","Type":"ContainerStarted","Data":"3e427856d41457853f68d5a32dc2f87168c0f52da592bc0a2b3db34f710fea2b"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.170036 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4nghl" event={"ID":"714cef39-2960-4a25-ac81-a4e65a115eb3","Type":"ContainerStarted","Data":"ea583be98e43e32e2b9eaab9f7b14cfcc6c03d18a9e1e7bc8ee858d8764bccfb"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.172116 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mrprd" event={"ID":"554abf87-b1ba-45b1-8130-95b40da3b8bf","Type":"ContainerStarted","Data":"e67c2cb59688caec6df23f5f60f848dd46db1badae2c1e268c4e6220c2473c3d"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.174456 4714 generic.go:334] "Generic (PLEG): container finished" podID="c779f8ba-7614-49f1-be6d-a9e316ec59ba" containerID="e783a0f04dc36f9b00a69d43ae0355c9b664ad354b4cdc74b2a094e5b287a951" exitCode=0 Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.174703 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" event={"ID":"c779f8ba-7614-49f1-be6d-a9e316ec59ba","Type":"ContainerDied","Data":"e783a0f04dc36f9b00a69d43ae0355c9b664ad354b4cdc74b2a094e5b287a951"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.177009 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ch6wr" event={"ID":"fc8e2d06-1cc2-4ea7-8d87-340d28740e20","Type":"ContainerStarted","Data":"06482b6470fadce059730244f351ad7eba0145dba4d03ba4db08d52aaec2d67b"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.181358 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-zkbcz" event={"ID":"f18250a8-66c1-445d-9452-081de13b24f7","Type":"ContainerStarted","Data":"da41d782933837b7976ce98d7f7dceca41fef963f862c55dd1a2cbca9a125e7e"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.182649 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-vhtdt" event={"ID":"02158b16-2eb1-4b8e-b1bb-55285b22d053","Type":"ContainerStarted","Data":"33677f709ea45fe339c09a2748d7e700229caf6e46e524c0ee7292bfc9a240d9"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.184583 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ljnh7" event={"ID":"38f843fe-c20b-4bc7-8f45-4ccd4b7be5a5","Type":"ContainerStarted","Data":"92cdec071830256f624ca0c36fdd787e417207e233c1063a8ab14f07ebc95d1f"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.186637 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vcj84" event={"ID":"eacb9f84-018a-4f64-b211-c9bedce50b9e","Type":"ContainerStarted","Data":"75c58bd7a88a96370485f93d6e2d9e0f9290e9433152047ce26bb9ea94c2b98d"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.188890 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-44gfk" event={"ID":"77b31235-8b07-4d66-aec8-64e5b7fae08e","Type":"ContainerStarted","Data":"b57b92cf1409311fba74582f0e99821728e24bd5d89e8468fe429886e3a64c84"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.204102 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" event={"ID":"8f71ba3e-c687-4ff7-9475-1e18ded764f6","Type":"ContainerStarted","Data":"fa826543db32457e92c9cbe60c432fd9da541a48d0e6863d88c03b0803b71041"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.205562 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jzg2m" event={"ID":"cc84f60e-094e-4924-b6f1-f0a8ab81aa4e","Type":"ContainerStarted","Data":"11084ed18f3b11987189819d9bb3dd492afdb82473bb2a69f517fe602c5a24f7"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.215758 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vcj84" podStartSLOduration=125.215738859 podStartE2EDuration="2m5.215738859s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:19.213224062 +0000 UTC m=+145.733725192" watchObservedRunningTime="2026-01-29 16:12:19.215738859 +0000 UTC m=+145.736239979" Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.221249 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chzp2" event={"ID":"1b632d84-c711-419a-9e24-bdb4c6e9aef6","Type":"ContainerStarted","Data":"2661bab2f985c589314989ca54a7b8417fe38b5661dd2093e413b43b2402ce34"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.224444 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt" event={"ID":"1d9869e2-6f55-4246-8ed0-b8af9dab3f74","Type":"ContainerStarted","Data":"55994348ab3dfc8cafc62c6e955628ce51995a40f12003eed7813e37443926ec"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.226628 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xtzbx" event={"ID":"2fa1ede8-3ea3-421d-929d-f6bf9cc1db0e","Type":"ContainerStarted","Data":"f4ea3f04681522c185b8823272251a5951ed23cf1e074da3e02d76c59ea7250d"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.228680 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sq9mx" event={"ID":"8062d225-aa57-48df-bf28-2254ecc4f635","Type":"ContainerStarted","Data":"951697dbf7c719b339005199178b35e78824e9709406b5beed446e6875f95f52"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.232630 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp5md" event={"ID":"0387af3d-8796-46b0-9282-9ecbda7fe3a7","Type":"ContainerStarted","Data":"39550e40b8e5b05c9cdd0a131397b8c7809c049c26f2073f6e73cf2dc88c7bb7"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.236262 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-z4h55" event={"ID":"bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92","Type":"ContainerStarted","Data":"92d27e916647ca2c73b1e771eece4264778a1470633157345c9df0fb0f9c40df"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.238666 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nf7jb" event={"ID":"2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96","Type":"ContainerStarted","Data":"d0ace5c4217a196b554ce0fd58847e9caf744ec08f66dd02e26fdb147fb683d2"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.248753 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" event={"ID":"80515d06-c09e-4c9d-a90f-43cc84edf4c9","Type":"ContainerStarted","Data":"5880f1855bae3fd6f603655d40b770623f038db9a3cb9db3918877f801567acc"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.257914 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj" event={"ID":"d288ee23-1753-48f2-ab82-736defe5fe18","Type":"ContainerStarted","Data":"3bfaf3a8d15fe9b3351c3d20e4177a0beef558e3311170e7b013daf0db5da97c"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.259585 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:19 crc kubenswrapper[4714]: E0129 16:12:19.261452 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:19.761421796 +0000 UTC m=+146.281922916 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.281885 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4dn69" event={"ID":"6de35940-bef4-4dfa-9a83-08ba29d73399","Type":"ContainerStarted","Data":"f9e1074728408def2a387a3b251afa2257fdc6dcfa08bf7a9f5de414199e7c75"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.284332 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" event={"ID":"2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab","Type":"ContainerStarted","Data":"bd5ddda0b956be079adf5a95904bc40a5bfab574b272bcc0aace47eca5063bd1"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.286348 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" event={"ID":"97cfbecd-36ef-409b-94e9-f607a1fa2c42","Type":"ContainerStarted","Data":"39b17c10dcdb044358f9409b55dce81fdd9556f01f697e8d12121445449b6325"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.287477 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8" event={"ID":"cb979c55-3027-4d92-94b9-cd17c32e6331","Type":"ContainerStarted","Data":"64dd0aa019d2ff78b021ae7336dcb6b797f2fc222a74d3ec8b1b6722bf60513e"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.288631 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-kfqcf" event={"ID":"706713ee-0ea2-4018-847c-ccf3a0fafb1c","Type":"ContainerStarted","Data":"ac69bf925281e62c711bc61e1e7af18c65e142666843c7ebd8119153dbb6c812"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.289393 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-jcdhl" event={"ID":"fcaee576-dff0-4a67-a0b1-7347b3030729","Type":"ContainerStarted","Data":"03a0c2abcdf4f7b3d6c47dcfce1a4e19bf09e285b4ddaa4cb1ab60d493e59e52"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.292279 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-zxb4v" event={"ID":"1fd5b799-74c2-4ffa-b3d9-6745c66ba28f","Type":"ContainerStarted","Data":"35d7262eb0de9b8a75316a252f9d4e76314213d335107d88f35442d4a8a41b55"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.294613 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-fn75b" event={"ID":"42b66dc3-a385-4350-a943-50f062da35f7","Type":"ContainerStarted","Data":"c4aaec06be7df88764d0dc745049e2e561eb871b6ccb463e86a9ef317a262a34"} Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.295453 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.298009 4714 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-xlczd container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.298045 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" podUID="3c2d0611-58f8-4a7e-8280-361c80d62802" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.314911 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-jb6jw" podStartSLOduration=125.314895681 podStartE2EDuration="2m5.314895681s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:19.311018483 +0000 UTC m=+145.831519603" watchObservedRunningTime="2026-01-29 16:12:19.314895681 +0000 UTC m=+145.835396801" Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.328938 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" podStartSLOduration=125.32891746 podStartE2EDuration="2m5.32891746s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:19.328230879 +0000 UTC m=+145.848731989" watchObservedRunningTime="2026-01-29 16:12:19.32891746 +0000 UTC m=+145.849418600" Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.345801 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-kvp9d" podStartSLOduration=125.345779846 podStartE2EDuration="2m5.345779846s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:19.343777014 +0000 UTC m=+145.864278134" watchObservedRunningTime="2026-01-29 16:12:19.345779846 +0000 UTC m=+145.866280966" Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.361052 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:19 crc kubenswrapper[4714]: E0129 16:12:19.362760 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:19.862739874 +0000 UTC m=+146.383241114 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.363591 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-m2g9h" podStartSLOduration=125.36357874 podStartE2EDuration="2m5.36357874s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:19.358008039 +0000 UTC m=+145.878509169" watchObservedRunningTime="2026-01-29 16:12:19.36357874 +0000 UTC m=+145.884079860" Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.461803 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:19 crc kubenswrapper[4714]: E0129 16:12:19.461969 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:19.961946838 +0000 UTC m=+146.482447958 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.462440 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:19 crc kubenswrapper[4714]: E0129 16:12:19.462747 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:19.962733482 +0000 UTC m=+146.483234602 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.563338 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:19 crc kubenswrapper[4714]: E0129 16:12:19.563705 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:20.063690729 +0000 UTC m=+146.584191849 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.664997 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:19 crc kubenswrapper[4714]: E0129 16:12:19.665485 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:20.165468981 +0000 UTC m=+146.685970101 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.766002 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:19 crc kubenswrapper[4714]: E0129 16:12:19.766623 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:20.266592223 +0000 UTC m=+146.787093393 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.867974 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:19 crc kubenswrapper[4714]: E0129 16:12:19.868333 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:20.368318483 +0000 UTC m=+146.888819593 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.969513 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:19 crc kubenswrapper[4714]: E0129 16:12:19.969731 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:20.469682573 +0000 UTC m=+146.990183733 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:19 crc kubenswrapper[4714]: I0129 16:12:19.969865 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:19 crc kubenswrapper[4714]: E0129 16:12:19.970312 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:20.470299222 +0000 UTC m=+146.990800342 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.071491 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:20 crc kubenswrapper[4714]: E0129 16:12:20.071803 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:20.571754694 +0000 UTC m=+147.092255874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.072081 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:20 crc kubenswrapper[4714]: E0129 16:12:20.072653 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:20.57262651 +0000 UTC m=+147.093127670 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.173500 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:20 crc kubenswrapper[4714]: E0129 16:12:20.173698 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:20.67366758 +0000 UTC m=+147.194168700 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.173894 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:20 crc kubenswrapper[4714]: E0129 16:12:20.174420 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:20.674399352 +0000 UTC m=+147.194900492 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.274791 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:20 crc kubenswrapper[4714]: E0129 16:12:20.275002 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:20.774973738 +0000 UTC m=+147.295474858 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.275391 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:20 crc kubenswrapper[4714]: E0129 16:12:20.275737 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:20.775729431 +0000 UTC m=+147.296230551 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.300094 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-sv7xw" event={"ID":"b7cf219f-4e80-47fc-b349-ea5c7eab6d9d","Type":"ContainerStarted","Data":"8a2fbce67ca39d9a948f00ef51d06e70b14ca049256c8ba6aa1a36330e72e95a"} Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.302055 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xvrxj" event={"ID":"f3e2f962-69e3-4008-a45f-5c35677f7f36","Type":"ContainerStarted","Data":"a45df571b922ec5a4ee61153c6edaae1fa7df4a72a624772ba8475e87c5aff9f"} Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.303976 4714 generic.go:334] "Generic (PLEG): container finished" podID="8f71ba3e-c687-4ff7-9475-1e18ded764f6" containerID="fa826543db32457e92c9cbe60c432fd9da541a48d0e6863d88c03b0803b71041" exitCode=0 Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.304077 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" event={"ID":"8f71ba3e-c687-4ff7-9475-1e18ded764f6","Type":"ContainerDied","Data":"fa826543db32457e92c9cbe60c432fd9da541a48d0e6863d88c03b0803b71041"} Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.307094 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jzg2m" event={"ID":"cc84f60e-094e-4924-b6f1-f0a8ab81aa4e","Type":"ContainerStarted","Data":"bd220881ff94f1948e3120bb7930942c7b879cd14685003c27e9c5391678f090"} Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.309055 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2lpzx" event={"ID":"c14fb55e-a42b-46c9-9521-6e8b60235166","Type":"ContainerStarted","Data":"a29d609a72c47494d9ab4a7af0ac3a2aa54fc3e5f33214e89b0bab625418fc8e"} Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.310787 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-lz6mw" event={"ID":"7a1dfb55-8680-4cbe-bd78-caca2e847caf","Type":"ContainerStarted","Data":"61719cba311aad8609ab840baea1cfa296842ec4a73281d02b334d955086776e"} Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.312486 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" event={"ID":"832097a5-4691-42b6-99cc-38679071d5ee","Type":"ContainerStarted","Data":"9bc1cd1aee2de10059a78dc94ebfc1cb6a64c8e3be39806bf6e6ab107f47abff"} Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.314695 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-99knh" event={"ID":"5b5da98c-0704-41c7-8563-707f7af93f41","Type":"ContainerStarted","Data":"581fea08a13cda1ef3ef4dbc47396d1c7474fc8fd004dcb57a27254dd7af9194"} Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.316421 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xtzbx" event={"ID":"2fa1ede8-3ea3-421d-929d-f6bf9cc1db0e","Type":"ContainerStarted","Data":"f0266f10504adfad4bf26dc8ced4f8847780aadf6b78ced8e81d31e0da10384f"} Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.318294 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" event={"ID":"80515d06-c09e-4c9d-a90f-43cc84edf4c9","Type":"ContainerStarted","Data":"2cfb1164c8a5f24d11bd2a23214b6a7408be50990447c790085e11ea6faaec50"} Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.320104 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt" event={"ID":"1d9869e2-6f55-4246-8ed0-b8af9dab3f74","Type":"ContainerStarted","Data":"b12fdb98d7913f55c00f8037769750024c6d7d15e2f1695d2fe08e49e4dde6c0"} Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.323198 4714 generic.go:334] "Generic (PLEG): container finished" podID="99bab267-639b-48b1-abc4-8c0373200a39" containerID="d74d43e34a1e11efc05f19a35eac3bfba23a056d00217dd74cd8226d7e7f07e3" exitCode=0 Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.323339 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6jl75" event={"ID":"99bab267-639b-48b1-abc4-8c0373200a39","Type":"ContainerDied","Data":"d74d43e34a1e11efc05f19a35eac3bfba23a056d00217dd74cd8226d7e7f07e3"} Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.323785 4714 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-xlczd container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.323826 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" podUID="3c2d0611-58f8-4a7e-8280-361c80d62802" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.324328 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-fn75b" Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.335597 4714 patch_prober.go:28] interesting pod/downloads-7954f5f757-fn75b container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.335674 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fn75b" podUID="42b66dc3-a385-4350-a943-50f062da35f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.358696 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" podStartSLOduration=125.358679767 podStartE2EDuration="2m5.358679767s" podCreationTimestamp="2026-01-29 16:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:20.357046337 +0000 UTC m=+146.877547457" watchObservedRunningTime="2026-01-29 16:12:20.358679767 +0000 UTC m=+146.879180877" Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.376809 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:20 crc kubenswrapper[4714]: E0129 16:12:20.377014 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:20.876993327 +0000 UTC m=+147.397494447 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.377430 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:20 crc kubenswrapper[4714]: E0129 16:12:20.379688 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:20.879669929 +0000 UTC m=+147.400171059 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.388969 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-fn75b" podStartSLOduration=126.388950003 podStartE2EDuration="2m6.388950003s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:20.388388116 +0000 UTC m=+146.908889246" watchObservedRunningTime="2026-01-29 16:12:20.388950003 +0000 UTC m=+146.909451123" Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.412964 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp5md" podStartSLOduration=126.412947176 podStartE2EDuration="2m6.412947176s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:20.412502363 +0000 UTC m=+146.933003503" watchObservedRunningTime="2026-01-29 16:12:20.412947176 +0000 UTC m=+146.933448296" Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.478146 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:20 crc kubenswrapper[4714]: E0129 16:12:20.478246 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:20.978230423 +0000 UTC m=+147.498731543 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:20 crc kubenswrapper[4714]: E0129 16:12:20.478776 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:20.978767999 +0000 UTC m=+147.499269119 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.478994 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.579724 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:20 crc kubenswrapper[4714]: E0129 16:12:20.580286 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:21.080266673 +0000 UTC m=+147.600767793 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.682095 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:20 crc kubenswrapper[4714]: E0129 16:12:20.682532 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:21.182517529 +0000 UTC m=+147.703018649 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.782868 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:20 crc kubenswrapper[4714]: E0129 16:12:20.783148 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:21.283120655 +0000 UTC m=+147.803621775 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.783237 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:20 crc kubenswrapper[4714]: E0129 16:12:20.783679 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:21.283667552 +0000 UTC m=+147.804168672 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.884010 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:20 crc kubenswrapper[4714]: E0129 16:12:20.884180 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:21.384157505 +0000 UTC m=+147.904658625 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.884363 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:20 crc kubenswrapper[4714]: E0129 16:12:20.884698 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:21.384685621 +0000 UTC m=+147.905186741 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.985719 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:20 crc kubenswrapper[4714]: E0129 16:12:20.985915 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:21.485884095 +0000 UTC m=+148.006385245 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:20 crc kubenswrapper[4714]: I0129 16:12:20.986456 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:20 crc kubenswrapper[4714]: E0129 16:12:20.986727 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:21.48671455 +0000 UTC m=+148.007215670 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.087806 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:21 crc kubenswrapper[4714]: E0129 16:12:21.088089 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:21.588055519 +0000 UTC m=+148.108556679 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.088336 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:21 crc kubenswrapper[4714]: E0129 16:12:21.088716 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:21.588700159 +0000 UTC m=+148.109201279 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.189210 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:21 crc kubenswrapper[4714]: E0129 16:12:21.189467 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:21.689431839 +0000 UTC m=+148.209932989 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.291519 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:21 crc kubenswrapper[4714]: E0129 16:12:21.292163 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:21.792136189 +0000 UTC m=+148.312637319 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.332124 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chzp2" event={"ID":"1b632d84-c711-419a-9e24-bdb4c6e9aef6","Type":"ContainerStarted","Data":"6bc694dad0122356b41ee216407b3a7bf24cc0a80090d885b685230aef3c95b4"} Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.333963 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sq9mx" event={"ID":"8062d225-aa57-48df-bf28-2254ecc4f635","Type":"ContainerStarted","Data":"dbb3fe5cab326b40963c30577a076031ca8d1773b098c85a11b511d73b7cab55"} Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.335864 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8" event={"ID":"cb979c55-3027-4d92-94b9-cd17c32e6331","Type":"ContainerStarted","Data":"8002880bca0bfedb17cd3285afc46c8ef8627aa8d830b5e5901fd7ddfd3b1e33"} Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.337683 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-44gfk" event={"ID":"77b31235-8b07-4d66-aec8-64e5b7fae08e","Type":"ContainerStarted","Data":"76badda10c81fd0945af3652cd492c0ee46a055ba17a3754657a75e06e651f6a"} Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.339073 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pbrmk" event={"ID":"adc3900b-dce0-4da4-bfc2-bca85b2395b2","Type":"ContainerStarted","Data":"c0dcbdcfc2435eacfaf4a331bec2b4bceb340d25fa4fa74db5426f3a18408104"} Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.340449 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-vhtdt" event={"ID":"02158b16-2eb1-4b8e-b1bb-55285b22d053","Type":"ContainerStarted","Data":"ece2bca6f5f907bdaaef3f0e5763d781791c7e86b194dbdc6b5211b4a92b9f39"} Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.341780 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ljnh7" event={"ID":"38f843fe-c20b-4bc7-8f45-4ccd4b7be5a5","Type":"ContainerStarted","Data":"12970cdc9701917e728084155eb6d878339e9df9344d7dd59010727b1bb27fcc"} Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.343277 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" event={"ID":"2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab","Type":"ContainerStarted","Data":"a3188f76f9132bb5a0195bb3f3bbc92354939921ea4aa61b37b491752b4801ba"} Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.344678 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-kfqcf" event={"ID":"706713ee-0ea2-4018-847c-ccf3a0fafb1c","Type":"ContainerStarted","Data":"3f153778ee3d6c5b312901d4605b400f6c87dd4cb70a25759341615692777c20"} Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.346092 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-zkbcz" event={"ID":"f18250a8-66c1-445d-9452-081de13b24f7","Type":"ContainerStarted","Data":"d51164f9bf54c0d21fc3ca36f0a0590a0e717bae11edde541ec1b0640a070ea6"} Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.347603 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-jcdhl" event={"ID":"fcaee576-dff0-4a67-a0b1-7347b3030729","Type":"ContainerStarted","Data":"ae8bbcc798871173e7ad0623699e7fd4462664f064c96ab7c3895b33345e3b8d"} Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.350153 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" event={"ID":"c779f8ba-7614-49f1-be6d-a9e316ec59ba","Type":"ContainerStarted","Data":"36273e1b01e65c4da1c39175bd578ab7ee8cf8f96e70f3732cc6c1016c11785a"} Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.351119 4714 patch_prober.go:28] interesting pod/downloads-7954f5f757-fn75b container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.351143 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.351166 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fn75b" podUID="42b66dc3-a385-4350-a943-50f062da35f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.352364 4714 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-l2t56 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.352398 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" podUID="80515d06-c09e-4c9d-a90f-43cc84edf4c9" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.376554 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-nh2m9" podStartSLOduration=127.37653799 podStartE2EDuration="2m7.37653799s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:21.375428016 +0000 UTC m=+147.895929136" watchObservedRunningTime="2026-01-29 16:12:21.37653799 +0000 UTC m=+147.897039110" Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.392794 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:21 crc kubenswrapper[4714]: E0129 16:12:21.393246 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:21.893204779 +0000 UTC m=+148.413705909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.393563 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:21 crc kubenswrapper[4714]: E0129 16:12:21.396801 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:21.896781269 +0000 UTC m=+148.417282599 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.400009 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-lz6mw" podStartSLOduration=127.399990727 podStartE2EDuration="2m7.399990727s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:21.398964366 +0000 UTC m=+147.919465496" watchObservedRunningTime="2026-01-29 16:12:21.399990727 +0000 UTC m=+147.920491847" Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.463151 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2lpzx" podStartSLOduration=127.463128038 podStartE2EDuration="2m7.463128038s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:21.461206419 +0000 UTC m=+147.981707539" watchObservedRunningTime="2026-01-29 16:12:21.463128038 +0000 UTC m=+147.983629158" Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.489187 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" podStartSLOduration=126.489167624 podStartE2EDuration="2m6.489167624s" podCreationTimestamp="2026-01-29 16:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:21.488751771 +0000 UTC m=+148.009252891" watchObservedRunningTime="2026-01-29 16:12:21.489167624 +0000 UTC m=+148.009668744" Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.497562 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:21 crc kubenswrapper[4714]: E0129 16:12:21.497834 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:21.997807708 +0000 UTC m=+148.518308828 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.546575 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mrprd" podStartSLOduration=127.546556919 podStartE2EDuration="2m7.546556919s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:21.527594209 +0000 UTC m=+148.048095329" watchObservedRunningTime="2026-01-29 16:12:21.546556919 +0000 UTC m=+148.067058029" Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.572885 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-zxb4v" podStartSLOduration=8.572866193 podStartE2EDuration="8.572866193s" podCreationTimestamp="2026-01-29 16:12:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:21.547141606 +0000 UTC m=+148.067642726" watchObservedRunningTime="2026-01-29 16:12:21.572866193 +0000 UTC m=+148.093367313" Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.573637 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" podStartSLOduration=127.573628876 podStartE2EDuration="2m7.573628876s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:21.570364847 +0000 UTC m=+148.090865967" watchObservedRunningTime="2026-01-29 16:12:21.573628876 +0000 UTC m=+148.094130006" Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.590426 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ch6wr" podStartSLOduration=126.590404589 podStartE2EDuration="2m6.590404589s" podCreationTimestamp="2026-01-29 16:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:21.589349317 +0000 UTC m=+148.109850437" watchObservedRunningTime="2026-01-29 16:12:21.590404589 +0000 UTC m=+148.110905709" Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.603677 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:21 crc kubenswrapper[4714]: E0129 16:12:21.604145 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:22.104131379 +0000 UTC m=+148.624632499 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.613679 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4dn69" podStartSLOduration=127.61363821 podStartE2EDuration="2m7.61363821s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:21.609671899 +0000 UTC m=+148.130173019" watchObservedRunningTime="2026-01-29 16:12:21.61363821 +0000 UTC m=+148.134139330" Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.704469 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:21 crc kubenswrapper[4714]: E0129 16:12:21.704754 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:22.204725385 +0000 UTC m=+148.725226525 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.704852 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:21 crc kubenswrapper[4714]: E0129 16:12:21.705244 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:22.205235001 +0000 UTC m=+148.725736111 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.738357 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.741174 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.741327 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.805720 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:21 crc kubenswrapper[4714]: E0129 16:12:21.805891 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:22.305863988 +0000 UTC m=+148.826365108 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.806141 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:21 crc kubenswrapper[4714]: E0129 16:12:21.806486 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:22.306474486 +0000 UTC m=+148.826975606 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.908083 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:21 crc kubenswrapper[4714]: E0129 16:12:21.908257 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:22.408233238 +0000 UTC m=+148.928734358 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:21 crc kubenswrapper[4714]: I0129 16:12:21.908373 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:21 crc kubenswrapper[4714]: E0129 16:12:21.908695 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:22.408686422 +0000 UTC m=+148.929187542 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.009735 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:22 crc kubenswrapper[4714]: E0129 16:12:22.009903 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:22.509879486 +0000 UTC m=+149.030380606 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.010091 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:22 crc kubenswrapper[4714]: E0129 16:12:22.010694 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:22.51066661 +0000 UTC m=+149.031167730 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.111054 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:22 crc kubenswrapper[4714]: E0129 16:12:22.111434 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:22.611419341 +0000 UTC m=+149.131920461 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.212307 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:22 crc kubenswrapper[4714]: E0129 16:12:22.212753 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:22.712734989 +0000 UTC m=+149.233236129 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.313657 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:22 crc kubenswrapper[4714]: E0129 16:12:22.313837 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:22.813810509 +0000 UTC m=+149.334311629 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.313968 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:22 crc kubenswrapper[4714]: E0129 16:12:22.314269 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:22.814257133 +0000 UTC m=+149.334758253 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.356676 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-99knh" event={"ID":"5b5da98c-0704-41c7-8563-707f7af93f41","Type":"ContainerStarted","Data":"fdfa177d70738b859553617206bc689e14aecd644fa4b96d1292a33aa982514a"} Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.358285 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj" event={"ID":"d288ee23-1753-48f2-ab82-736defe5fe18","Type":"ContainerStarted","Data":"5217358d817ce7eff91c5ce92675d69f303d8fb317d488bdfcb19f7a4a81e957"} Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.360744 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6jl75" event={"ID":"99bab267-639b-48b1-abc4-8c0373200a39","Type":"ContainerStarted","Data":"0649707319cf315cb884412df9aee13b92a185a2ed35bc0133278b8e0ac6b87a"} Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.362431 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt" event={"ID":"1d9869e2-6f55-4246-8ed0-b8af9dab3f74","Type":"ContainerStarted","Data":"20a2b239f390c2ee891200c241fcd960b389bf031f732ea731cef1b393235d6f"} Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.368701 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-z4h55" event={"ID":"bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92","Type":"ContainerStarted","Data":"e62dd24fddcd7bdb174cdd4cfdba6907f75f804fea4a0898502c6d9dac1157bd"} Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.371184 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-sv7xw" event={"ID":"b7cf219f-4e80-47fc-b349-ea5c7eab6d9d","Type":"ContainerStarted","Data":"e19b79c6c66557c4fe43cea947e82f0138aa0b8582b2b23582a033be4c765462"} Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.372847 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nf7jb" event={"ID":"2eab9b06-06e4-4f58-ab86-ab1bf3b5cc96","Type":"ContainerStarted","Data":"4a8a76e56d6dacf387ad607f5c069cb41422dd571ddbdedad5ba541c251edfae"} Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.375578 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" event={"ID":"8f71ba3e-c687-4ff7-9475-1e18ded764f6","Type":"ContainerStarted","Data":"0280359fe4f163eeaed24796981149340b0af92903e6c783ada39027df0eb8e3"} Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.376480 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.376489 4714 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-l2t56 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.376549 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" podUID="80515d06-c09e-4c9d-a90f-43cc84edf4c9" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.380061 4714 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-v68nn container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.380117 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" podUID="2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.395095 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9thpj" podStartSLOduration=128.395074014 podStartE2EDuration="2m8.395074014s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:22.374721932 +0000 UTC m=+148.895223052" watchObservedRunningTime="2026-01-29 16:12:22.395074014 +0000 UTC m=+148.915575134" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.415662 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-kfqcf" podStartSLOduration=127.415645223 podStartE2EDuration="2m7.415645223s" podCreationTimestamp="2026-01-29 16:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:22.41490972 +0000 UTC m=+148.935410850" watchObservedRunningTime="2026-01-29 16:12:22.415645223 +0000 UTC m=+148.936146343" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.417401 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-z4h55" podStartSLOduration=127.417390406 podStartE2EDuration="2m7.417390406s" podCreationTimestamp="2026-01-29 16:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:22.397641632 +0000 UTC m=+148.918142762" watchObservedRunningTime="2026-01-29 16:12:22.417390406 +0000 UTC m=+148.937891526" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.426196 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:22 crc kubenswrapper[4714]: E0129 16:12:22.426407 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:22.926379291 +0000 UTC m=+149.446880411 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.426590 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.426631 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.426657 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.426701 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.426998 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:12:22 crc kubenswrapper[4714]: E0129 16:12:22.432670 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:22.932650423 +0000 UTC m=+149.453151623 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.434632 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.435577 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.436902 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.439730 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8" podStartSLOduration=127.439720769 podStartE2EDuration="2m7.439720769s" podCreationTimestamp="2026-01-29 16:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:22.438238924 +0000 UTC m=+148.958740054" watchObservedRunningTime="2026-01-29 16:12:22.439720769 +0000 UTC m=+148.960221879" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.441177 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.468063 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ljnh7" podStartSLOduration=127.468043795 podStartE2EDuration="2m7.468043795s" podCreationTimestamp="2026-01-29 16:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:22.45281909 +0000 UTC m=+148.973320240" watchObservedRunningTime="2026-01-29 16:12:22.468043795 +0000 UTC m=+148.988544915" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.470004 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-chzp2" podStartSLOduration=128.469989885 podStartE2EDuration="2m8.469989885s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:22.466998913 +0000 UTC m=+148.987500033" watchObservedRunningTime="2026-01-29 16:12:22.469989885 +0000 UTC m=+148.990491015" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.483374 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" podStartSLOduration=128.483356713 podStartE2EDuration="2m8.483356713s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:22.481534108 +0000 UTC m=+149.002035228" watchObservedRunningTime="2026-01-29 16:12:22.483356713 +0000 UTC m=+149.003857833" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.496326 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-44gfk" podStartSLOduration=127.49631164 podStartE2EDuration="2m7.49631164s" podCreationTimestamp="2026-01-29 16:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:22.49502951 +0000 UTC m=+149.015530630" watchObservedRunningTime="2026-01-29 16:12:22.49631164 +0000 UTC m=+149.016812760" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.508875 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-vhtdt" podStartSLOduration=9.508853643 podStartE2EDuration="9.508853643s" podCreationTimestamp="2026-01-29 16:12:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:22.507551433 +0000 UTC m=+149.028052553" watchObservedRunningTime="2026-01-29 16:12:22.508853643 +0000 UTC m=+149.029354763" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.519124 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.524409 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.525312 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" podStartSLOduration=127.525292666 podStartE2EDuration="2m7.525292666s" podCreationTimestamp="2026-01-29 16:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:22.524025847 +0000 UTC m=+149.044526967" watchObservedRunningTime="2026-01-29 16:12:22.525292666 +0000 UTC m=+149.045793796" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.527411 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:22 crc kubenswrapper[4714]: E0129 16:12:22.527848 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:23.027812733 +0000 UTC m=+149.548313863 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.552261 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sq9mx" podStartSLOduration=127.55224134 podStartE2EDuration="2m7.55224134s" podCreationTimestamp="2026-01-29 16:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:22.551034893 +0000 UTC m=+149.071536013" watchObservedRunningTime="2026-01-29 16:12:22.55224134 +0000 UTC m=+149.072742460" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.573414 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xvrxj" podStartSLOduration=128.573397327 podStartE2EDuration="2m8.573397327s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:22.572275802 +0000 UTC m=+149.092776922" watchObservedRunningTime="2026-01-29 16:12:22.573397327 +0000 UTC m=+149.093898457" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.599044 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.599919 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" podStartSLOduration=127.599908237 podStartE2EDuration="2m7.599908237s" podCreationTimestamp="2026-01-29 16:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:22.599077972 +0000 UTC m=+149.119579092" watchObservedRunningTime="2026-01-29 16:12:22.599908237 +0000 UTC m=+149.120409357" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.621195 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jzg2m" podStartSLOduration=127.621177328 podStartE2EDuration="2m7.621177328s" podCreationTimestamp="2026-01-29 16:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:22.620821987 +0000 UTC m=+149.141323107" watchObservedRunningTime="2026-01-29 16:12:22.621177328 +0000 UTC m=+149.141678458" Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.629078 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:22 crc kubenswrapper[4714]: E0129 16:12:22.629366 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:23.129355698 +0000 UTC m=+149.649856818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.732727 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:22 crc kubenswrapper[4714]: E0129 16:12:22.733341 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:23.233325436 +0000 UTC m=+149.753826556 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.750502 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.750556 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Jan 29 16:12:22 crc kubenswrapper[4714]: W0129 16:12:22.791208 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-2f2774f8508cbdbf6b13c5b366a676e0f00b06214da0c20f702f75031ddf6782 WatchSource:0}: Error finding container 2f2774f8508cbdbf6b13c5b366a676e0f00b06214da0c20f702f75031ddf6782: Status 404 returned error can't find the container with id 2f2774f8508cbdbf6b13c5b366a676e0f00b06214da0c20f702f75031ddf6782 Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.836480 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:22 crc kubenswrapper[4714]: E0129 16:12:22.836860 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:23.336847471 +0000 UTC m=+149.857348581 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:22 crc kubenswrapper[4714]: W0129 16:12:22.893822 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-233de5020fb91da4466eb92d9dbe85fa38ae483b312d7f505f7c7f9fcc6c97ae WatchSource:0}: Error finding container 233de5020fb91da4466eb92d9dbe85fa38ae483b312d7f505f7c7f9fcc6c97ae: Status 404 returned error can't find the container with id 233de5020fb91da4466eb92d9dbe85fa38ae483b312d7f505f7c7f9fcc6c97ae Jan 29 16:12:22 crc kubenswrapper[4714]: I0129 16:12:22.937790 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:22 crc kubenswrapper[4714]: E0129 16:12:22.938166 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:23.438148109 +0000 UTC m=+149.958649229 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.040370 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:23 crc kubenswrapper[4714]: E0129 16:12:23.040637 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:23.540625422 +0000 UTC m=+150.061126542 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.141060 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:23 crc kubenswrapper[4714]: E0129 16:12:23.141247 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:23.641219498 +0000 UTC m=+150.161720628 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.141321 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:23 crc kubenswrapper[4714]: E0129 16:12:23.141607 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:23.64159795 +0000 UTC m=+150.162099080 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.242524 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:23 crc kubenswrapper[4714]: E0129 16:12:23.242708 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:23.74267722 +0000 UTC m=+150.263178350 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.242784 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:23 crc kubenswrapper[4714]: E0129 16:12:23.243122 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:23.743110364 +0000 UTC m=+150.263611474 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.344458 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:23 crc kubenswrapper[4714]: E0129 16:12:23.344702 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:23.844671979 +0000 UTC m=+150.365173099 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.345018 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:23 crc kubenswrapper[4714]: E0129 16:12:23.345403 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:23.845388671 +0000 UTC m=+150.365889801 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.383147 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"233de5020fb91da4466eb92d9dbe85fa38ae483b312d7f505f7c7f9fcc6c97ae"} Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.384523 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"803bf00dc059eba74bdf6347eeb2d3bed51ce9ca6e5847b0b7980ffcfe2f8519"} Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.385843 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"31d4d8de5b9e812dda05d91762b6d773fa399a230b3ec3fce86d9285fe3fbdc3"} Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.385897 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"2f2774f8508cbdbf6b13c5b366a676e0f00b06214da0c20f702f75031ddf6782"} Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.388116 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6jl75" event={"ID":"99bab267-639b-48b1-abc4-8c0373200a39","Type":"ContainerStarted","Data":"16e55a14fb4aaaa9b4a2ff320e0057b8299dcd5719d4e77935929d031851eb87"} Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.390209 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xtzbx" event={"ID":"2fa1ede8-3ea3-421d-929d-f6bf9cc1db0e","Type":"ContainerStarted","Data":"c5f70c46d50b5294e7f36701d45692d9d1f41d9d50460dd007c421dba9653d17"} Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.392524 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-zkbcz" event={"ID":"f18250a8-66c1-445d-9452-081de13b24f7","Type":"ContainerStarted","Data":"c5f98c48f6813667c5fbd018311042683d511aa37bfde013ffde64453733828b"} Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.395072 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-jcdhl" event={"ID":"fcaee576-dff0-4a67-a0b1-7347b3030729","Type":"ContainerStarted","Data":"464ad6a3b23bf3d892319405739fd7bdedc4a42ece626b872203487f1c929e1a"} Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.395199 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-jcdhl" Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.400459 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pbrmk" event={"ID":"adc3900b-dce0-4da4-bfc2-bca85b2395b2","Type":"ContainerStarted","Data":"db7e47982d58f7123b5f33d45571e2ab833e4d49905865e79880a686399644fb"} Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.400510 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pbrmk" Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.400882 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.403304 4714 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-v68nn container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.403360 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" podUID="2fbdedcd-e0ce-4fe9-b19a-53d422b1fcab" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.408731 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xtzbx" podStartSLOduration=128.408715227 podStartE2EDuration="2m8.408715227s" podCreationTimestamp="2026-01-29 16:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:23.408234873 +0000 UTC m=+149.928736013" watchObservedRunningTime="2026-01-29 16:12:23.408715227 +0000 UTC m=+149.929216357" Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.433683 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nf7jb" podStartSLOduration=129.43366403 podStartE2EDuration="2m9.43366403s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:23.430810183 +0000 UTC m=+149.951311303" watchObservedRunningTime="2026-01-29 16:12:23.43366403 +0000 UTC m=+149.954165170" Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.446617 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:23 crc kubenswrapper[4714]: E0129 16:12:23.447268 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:23.947246585 +0000 UTC m=+150.467747715 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.447925 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:23 crc kubenswrapper[4714]: E0129 16:12:23.450005 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:23.949988949 +0000 UTC m=+150.470490079 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.489047 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-sv7xw" podStartSLOduration=129.489029753 podStartE2EDuration="2m9.489029753s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:23.463825622 +0000 UTC m=+149.984326742" watchObservedRunningTime="2026-01-29 16:12:23.489029753 +0000 UTC m=+150.009530873" Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.490883 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-jcdhl" podStartSLOduration=10.490874629 podStartE2EDuration="10.490874629s" podCreationTimestamp="2026-01-29 16:12:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:23.488412144 +0000 UTC m=+150.008913264" watchObservedRunningTime="2026-01-29 16:12:23.490874629 +0000 UTC m=+150.011375749" Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.506976 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r6cxt" podStartSLOduration=129.506948851 podStartE2EDuration="2m9.506948851s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:23.504730593 +0000 UTC m=+150.025231713" watchObservedRunningTime="2026-01-29 16:12:23.506948851 +0000 UTC m=+150.027449971" Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.529224 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-99knh" podStartSLOduration=129.529208522 podStartE2EDuration="2m9.529208522s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:23.525700324 +0000 UTC m=+150.046201444" watchObservedRunningTime="2026-01-29 16:12:23.529208522 +0000 UTC m=+150.049709642" Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.548973 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:23 crc kubenswrapper[4714]: E0129 16:12:23.549308 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:24.049292736 +0000 UTC m=+150.569793856 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.559617 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-zkbcz" podStartSLOduration=128.559603881 podStartE2EDuration="2m8.559603881s" podCreationTimestamp="2026-01-29 16:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:23.558321272 +0000 UTC m=+150.078822392" watchObservedRunningTime="2026-01-29 16:12:23.559603881 +0000 UTC m=+150.080105001" Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.650450 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:23 crc kubenswrapper[4714]: E0129 16:12:23.650832 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:24.15081544 +0000 UTC m=+150.671316560 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.740990 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 16:12:23 crc kubenswrapper[4714]: [-]has-synced failed: reason withheld Jan 29 16:12:23 crc kubenswrapper[4714]: [+]process-running ok Jan 29 16:12:23 crc kubenswrapper[4714]: healthz check failed Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.741064 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.751824 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:23 crc kubenswrapper[4714]: E0129 16:12:23.752177 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:24.252145328 +0000 UTC m=+150.772646448 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.752262 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:23 crc kubenswrapper[4714]: E0129 16:12:23.752643 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:24.252630823 +0000 UTC m=+150.773131943 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.853263 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:23 crc kubenswrapper[4714]: E0129 16:12:23.853412 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:24.353391584 +0000 UTC m=+150.873892704 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.853448 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:23 crc kubenswrapper[4714]: E0129 16:12:23.853759 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:24.353751475 +0000 UTC m=+150.874252595 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.954531 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:23 crc kubenswrapper[4714]: E0129 16:12:23.954720 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:24.454698442 +0000 UTC m=+150.975199562 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:23 crc kubenswrapper[4714]: I0129 16:12:23.954780 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:23 crc kubenswrapper[4714]: E0129 16:12:23.955118 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:24.455111074 +0000 UTC m=+150.975612194 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.055809 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:24 crc kubenswrapper[4714]: E0129 16:12:24.056034 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:24.5560084 +0000 UTC m=+151.076509520 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.056420 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:24 crc kubenswrapper[4714]: E0129 16:12:24.056719 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:24.556709811 +0000 UTC m=+151.077210931 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.157324 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:24 crc kubenswrapper[4714]: E0129 16:12:24.157464 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:24.657440401 +0000 UTC m=+151.177941521 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.157511 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:24 crc kubenswrapper[4714]: E0129 16:12:24.157842 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:24.657827893 +0000 UTC m=+151.178329053 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.205736 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pbrmk" podStartSLOduration=129.205700887 podStartE2EDuration="2m9.205700887s" podCreationTimestamp="2026-01-29 16:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:23.578906611 +0000 UTC m=+150.099407731" watchObservedRunningTime="2026-01-29 16:12:24.205700887 +0000 UTC m=+150.726202007" Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.259100 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:24 crc kubenswrapper[4714]: E0129 16:12:24.259282 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:24.759252914 +0000 UTC m=+151.279754034 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.259368 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:24 crc kubenswrapper[4714]: E0129 16:12:24.259730 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:24.759720389 +0000 UTC m=+151.280221589 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.360339 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:24 crc kubenswrapper[4714]: E0129 16:12:24.360571 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:24.860522791 +0000 UTC m=+151.381023921 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.360647 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:24 crc kubenswrapper[4714]: E0129 16:12:24.361234 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:24.861219562 +0000 UTC m=+151.381720692 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.376505 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.378663 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.385586 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.385873 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.390772 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.426519 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7c90fd8a3db23778d0ce37dffb6807206e355df3febeeeab78212b0d154f7f50"} Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.429203 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"59f2ceae6c23231edc9f1b9ef386c9a7f144a70e49fa01817a311892266720b9"} Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.452083 4714 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-dwsm5 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.452146 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" podUID="8f71ba3e-c687-4ff7-9475-1e18ded764f6" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.463409 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.463690 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fee5819c-8349-4080-9922-453f31a300da-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fee5819c-8349-4080-9922-453f31a300da\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.463878 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fee5819c-8349-4080-9922-453f31a300da-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fee5819c-8349-4080-9922-453f31a300da\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 16:12:24 crc kubenswrapper[4714]: E0129 16:12:24.465185 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:24.96516058 +0000 UTC m=+151.485661730 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.499439 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-6jl75" podStartSLOduration=130.499421798 podStartE2EDuration="2m10.499421798s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:24.497885391 +0000 UTC m=+151.018386511" watchObservedRunningTime="2026-01-29 16:12:24.499421798 +0000 UTC m=+151.019922918" Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.565150 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.565221 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fee5819c-8349-4080-9922-453f31a300da-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fee5819c-8349-4080-9922-453f31a300da\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.565290 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fee5819c-8349-4080-9922-453f31a300da-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fee5819c-8349-4080-9922-453f31a300da\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 16:12:24 crc kubenswrapper[4714]: E0129 16:12:24.565865 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:25.065849619 +0000 UTC m=+151.586350749 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.566068 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fee5819c-8349-4080-9922-453f31a300da-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fee5819c-8349-4080-9922-453f31a300da\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.608786 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fee5819c-8349-4080-9922-453f31a300da-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fee5819c-8349-4080-9922-453f31a300da\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.666174 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:24 crc kubenswrapper[4714]: E0129 16:12:24.666612 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:25.16659361 +0000 UTC m=+151.687094730 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.706105 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.741239 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 16:12:24 crc kubenswrapper[4714]: [-]has-synced failed: reason withheld Jan 29 16:12:24 crc kubenswrapper[4714]: [+]process-running ok Jan 29 16:12:24 crc kubenswrapper[4714]: healthz check failed Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.741284 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.768161 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:24 crc kubenswrapper[4714]: E0129 16:12:24.768436 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:25.268425353 +0000 UTC m=+151.788926473 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.778058 4714 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-dwsm5 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.778096 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" podUID="8f71ba3e-c687-4ff7-9475-1e18ded764f6" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.778132 4714 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-dwsm5 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.778195 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" podUID="8f71ba3e-c687-4ff7-9475-1e18ded764f6" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.869723 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:24 crc kubenswrapper[4714]: E0129 16:12:24.870091 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:25.370076581 +0000 UTC m=+151.890577701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:24 crc kubenswrapper[4714]: I0129 16:12:24.971248 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:24 crc kubenswrapper[4714]: E0129 16:12:24.971580 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:25.471568875 +0000 UTC m=+151.992069995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.021523 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.072239 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:25 crc kubenswrapper[4714]: E0129 16:12:25.072610 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:25.572594774 +0000 UTC m=+152.093095894 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.173795 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:25 crc kubenswrapper[4714]: E0129 16:12:25.174253 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:25.674239152 +0000 UTC m=+152.194740272 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.275367 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:25 crc kubenswrapper[4714]: E0129 16:12:25.275536 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:25.775507858 +0000 UTC m=+152.296008978 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.275829 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:25 crc kubenswrapper[4714]: E0129 16:12:25.276146 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:25.776136668 +0000 UTC m=+152.296637858 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.377158 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:25 crc kubenswrapper[4714]: E0129 16:12:25.377433 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:25.877419275 +0000 UTC m=+152.397920395 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.433503 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4nghl" event={"ID":"714cef39-2960-4a25-ac81-a4e65a115eb3","Type":"ContainerStarted","Data":"94e108c84782e87694fda57ab7922c324fb5c9f04546542dd16572b2af15e6d5"} Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.435112 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"fee5819c-8349-4080-9922-453f31a300da","Type":"ContainerStarted","Data":"aabce909ed74e646558bee9fa02988139c96b4993b0d41ddfcd7351dba0ff624"} Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.478794 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:25 crc kubenswrapper[4714]: E0129 16:12:25.479411 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:25.979395633 +0000 UTC m=+152.499896753 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.511702 4714 csr.go:261] certificate signing request csr-hc25h is approved, waiting to be issued Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.519171 4714 csr.go:257] certificate signing request csr-hc25h is issued Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.579531 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:25 crc kubenswrapper[4714]: E0129 16:12:25.579714 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:26.079656508 +0000 UTC m=+152.600157628 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.579770 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:25 crc kubenswrapper[4714]: E0129 16:12:25.580095 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:26.080087582 +0000 UTC m=+152.600588702 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.607982 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-kvp9d" Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.612979 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-kvp9d" Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.680650 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:25 crc kubenswrapper[4714]: E0129 16:12:25.680862 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:26.180841862 +0000 UTC m=+152.701342982 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.680911 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.680967 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.681140 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:25 crc kubenswrapper[4714]: E0129 16:12:25.681459 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:26.181450301 +0000 UTC m=+152.701951421 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.682665 4714 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-kgl5s container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.8:8443/livez\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.682716 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" podUID="c779f8ba-7614-49f1-be6d-a9e316ec59ba" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.8:8443/livez\": dial tcp 10.217.0.8:8443: connect: connection refused" Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.716927 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.716997 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.718890 4714 patch_prober.go:28] interesting pod/console-f9d7485db-m2g9h container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.718943 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-m2g9h" podUID="0e2a789d-6a90-4d60-881e-9562cd92e0a7" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.741588 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 16:12:25 crc kubenswrapper[4714]: [-]has-synced failed: reason withheld Jan 29 16:12:25 crc kubenswrapper[4714]: [+]process-running ok Jan 29 16:12:25 crc kubenswrapper[4714]: healthz check failed Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.741643 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.782102 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:25 crc kubenswrapper[4714]: E0129 16:12:25.782285 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:26.282267034 +0000 UTC m=+152.802768154 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.782344 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:25 crc kubenswrapper[4714]: E0129 16:12:25.782781 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:26.282764579 +0000 UTC m=+152.803265699 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.795250 4714 patch_prober.go:28] interesting pod/downloads-7954f5f757-fn75b container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.795260 4714 patch_prober.go:28] interesting pod/downloads-7954f5f757-fn75b container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.795354 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-fn75b" podUID="42b66dc3-a385-4350-a943-50f062da35f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.795309 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fn75b" podUID="42b66dc3-a385-4350-a943-50f062da35f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.877413 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.884660 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:25 crc kubenswrapper[4714]: E0129 16:12:25.885003 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:26.384988925 +0000 UTC m=+152.905490045 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.910347 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.920806 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.986476 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:25 crc kubenswrapper[4714]: E0129 16:12:25.987449 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:26.487432777 +0000 UTC m=+153.007933897 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.992224 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.992261 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.993173 4714 patch_prober.go:28] interesting pod/apiserver-76f77b778f-6jl75 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.10:8443/livez\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Jan 29 16:12:25 crc kubenswrapper[4714]: I0129 16:12:25.993210 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-6jl75" podUID="99bab267-639b-48b1-abc4-8c0373200a39" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.10:8443/livez\": dial tcp 10.217.0.10:8443: connect: connection refused" Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.087393 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:26 crc kubenswrapper[4714]: E0129 16:12:26.087560 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:26.587536368 +0000 UTC m=+153.108037488 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.087635 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:26 crc kubenswrapper[4714]: E0129 16:12:26.088073 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:26.588053764 +0000 UTC m=+153.108554884 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.101786 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.110982 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.188726 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:26 crc kubenswrapper[4714]: E0129 16:12:26.188898 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:26.688862066 +0000 UTC m=+153.209363186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.189074 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:26 crc kubenswrapper[4714]: E0129 16:12:26.190271 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:26.690257999 +0000 UTC m=+153.210759119 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.289895 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:26 crc kubenswrapper[4714]: E0129 16:12:26.290090 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:26.79006323 +0000 UTC m=+153.310564350 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.290155 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:26 crc kubenswrapper[4714]: E0129 16:12:26.290490 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:26.790482692 +0000 UTC m=+153.310983812 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.391400 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:26 crc kubenswrapper[4714]: E0129 16:12:26.391590 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:26.891559763 +0000 UTC m=+153.412060883 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.391638 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:26 crc kubenswrapper[4714]: E0129 16:12:26.391966 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:26.891952565 +0000 UTC m=+153.412453685 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.442484 4714 generic.go:334] "Generic (PLEG): container finished" podID="cb979c55-3027-4d92-94b9-cd17c32e6331" containerID="8002880bca0bfedb17cd3285afc46c8ef8627aa8d830b5e5901fd7ddfd3b1e33" exitCode=0 Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.442571 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8" event={"ID":"cb979c55-3027-4d92-94b9-cd17c32e6331","Type":"ContainerDied","Data":"8002880bca0bfedb17cd3285afc46c8ef8627aa8d830b5e5901fd7ddfd3b1e33"} Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.445182 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"fee5819c-8349-4080-9922-453f31a300da","Type":"ContainerStarted","Data":"35db7973ed160da110c671870acde50bc58a4cf360b02511f01c25b5e19a9de9"} Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.489747 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.489727385 podStartE2EDuration="2.489727385s" podCreationTimestamp="2026-01-29 16:12:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:26.488850638 +0000 UTC m=+153.009351758" watchObservedRunningTime="2026-01-29 16:12:26.489727385 +0000 UTC m=+153.010228505" Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.492564 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:26 crc kubenswrapper[4714]: E0129 16:12:26.492890 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:26.99285198 +0000 UTC m=+153.513353100 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.492989 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:26 crc kubenswrapper[4714]: E0129 16:12:26.494251 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:26.994234993 +0000 UTC m=+153.514736113 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.513579 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jzg2m" Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.519768 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jzg2m" Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.520083 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-01-29 16:07:25 +0000 UTC, rotation deadline is 2026-10-30 04:21:57.88643458 +0000 UTC Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.520133 4714 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6564h9m31.366304766s for next certificate rotation Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.594153 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:26 crc kubenswrapper[4714]: E0129 16:12:26.595333 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:27.095313483 +0000 UTC m=+153.615814603 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.695858 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:26 crc kubenswrapper[4714]: E0129 16:12:26.696310 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:27.196287921 +0000 UTC m=+153.716789041 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.738018 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.744007 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xtr82"] Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.745180 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xtr82" Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.745891 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 16:12:26 crc kubenswrapper[4714]: [-]has-synced failed: reason withheld Jan 29 16:12:26 crc kubenswrapper[4714]: [+]process-running ok Jan 29 16:12:26 crc kubenswrapper[4714]: healthz check failed Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.745962 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.758852 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.791358 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.797456 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:26 crc kubenswrapper[4714]: E0129 16:12:26.797653 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:27.297621969 +0000 UTC m=+153.818123089 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.797709 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbn22\" (UniqueName: \"kubernetes.io/projected/11a30de8-b234-47b4-8fd0-44f0c428be78-kube-api-access-zbn22\") pod \"community-operators-xtr82\" (UID: \"11a30de8-b234-47b4-8fd0-44f0c428be78\") " pod="openshift-marketplace/community-operators-xtr82" Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.797814 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11a30de8-b234-47b4-8fd0-44f0c428be78-utilities\") pod \"community-operators-xtr82\" (UID: \"11a30de8-b234-47b4-8fd0-44f0c428be78\") " pod="openshift-marketplace/community-operators-xtr82" Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.798057 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.798249 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11a30de8-b234-47b4-8fd0-44f0c428be78-catalog-content\") pod \"community-operators-xtr82\" (UID: \"11a30de8-b234-47b4-8fd0-44f0c428be78\") " pod="openshift-marketplace/community-operators-xtr82" Jan 29 16:12:26 crc kubenswrapper[4714]: E0129 16:12:26.798455 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:27.298440334 +0000 UTC m=+153.818941454 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.815373 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xtr82"] Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.895685 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ljnh7" Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.899976 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.900340 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11a30de8-b234-47b4-8fd0-44f0c428be78-catalog-content\") pod \"community-operators-xtr82\" (UID: \"11a30de8-b234-47b4-8fd0-44f0c428be78\") " pod="openshift-marketplace/community-operators-xtr82" Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.900373 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbn22\" (UniqueName: \"kubernetes.io/projected/11a30de8-b234-47b4-8fd0-44f0c428be78-kube-api-access-zbn22\") pod \"community-operators-xtr82\" (UID: \"11a30de8-b234-47b4-8fd0-44f0c428be78\") " pod="openshift-marketplace/community-operators-xtr82" Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.900403 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11a30de8-b234-47b4-8fd0-44f0c428be78-utilities\") pod \"community-operators-xtr82\" (UID: \"11a30de8-b234-47b4-8fd0-44f0c428be78\") " pod="openshift-marketplace/community-operators-xtr82" Jan 29 16:12:26 crc kubenswrapper[4714]: E0129 16:12:26.901428 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:27.401410383 +0000 UTC m=+153.921911503 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.901918 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11a30de8-b234-47b4-8fd0-44f0c428be78-catalog-content\") pod \"community-operators-xtr82\" (UID: \"11a30de8-b234-47b4-8fd0-44f0c428be78\") " pod="openshift-marketplace/community-operators-xtr82" Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.902068 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11a30de8-b234-47b4-8fd0-44f0c428be78-utilities\") pod \"community-operators-xtr82\" (UID: \"11a30de8-b234-47b4-8fd0-44f0c428be78\") " pod="openshift-marketplace/community-operators-xtr82" Jan 29 16:12:26 crc kubenswrapper[4714]: I0129 16:12:26.910327 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ljnh7" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.002656 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:27 crc kubenswrapper[4714]: E0129 16:12:27.003313 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:27.503302038 +0000 UTC m=+154.023803158 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.005759 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbn22\" (UniqueName: \"kubernetes.io/projected/11a30de8-b234-47b4-8fd0-44f0c428be78-kube-api-access-zbn22\") pod \"community-operators-xtr82\" (UID: \"11a30de8-b234-47b4-8fd0-44f0c428be78\") " pod="openshift-marketplace/community-operators-xtr82" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.023039 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-74twj"] Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.023904 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-74twj" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.040589 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.056736 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.058012 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.061252 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xtr82" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.085578 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.092256 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.110565 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-74twj"] Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.113399 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.113670 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a97ed1ff-657f-4bde-943b-78caf9d07f92-utilities\") pod \"certified-operators-74twj\" (UID: \"a97ed1ff-657f-4bde-943b-78caf9d07f92\") " pod="openshift-marketplace/certified-operators-74twj" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.113709 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.113786 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.113803 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8bfv\" (UniqueName: \"kubernetes.io/projected/a97ed1ff-657f-4bde-943b-78caf9d07f92-kube-api-access-v8bfv\") pod \"certified-operators-74twj\" (UID: \"a97ed1ff-657f-4bde-943b-78caf9d07f92\") " pod="openshift-marketplace/certified-operators-74twj" Jan 29 16:12:27 crc kubenswrapper[4714]: E0129 16:12:27.113867 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:27.613836838 +0000 UTC m=+154.134337958 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.113985 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a97ed1ff-657f-4bde-943b-78caf9d07f92-catalog-content\") pod \"certified-operators-74twj\" (UID: \"a97ed1ff-657f-4bde-943b-78caf9d07f92\") " pod="openshift-marketplace/certified-operators-74twj" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.145669 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.166368 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v68nn" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.168607 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6bjgq"] Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.169867 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6bjgq" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.213980 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6bjgq"] Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.215038 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.215100 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98a35d03-ef3b-4341-9866-56d12a28aee3-utilities\") pod \"community-operators-6bjgq\" (UID: \"98a35d03-ef3b-4341-9866-56d12a28aee3\") " pod="openshift-marketplace/community-operators-6bjgq" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.215162 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.215188 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gks4z\" (UniqueName: \"kubernetes.io/projected/98a35d03-ef3b-4341-9866-56d12a28aee3-kube-api-access-gks4z\") pod \"community-operators-6bjgq\" (UID: \"98a35d03-ef3b-4341-9866-56d12a28aee3\") " pod="openshift-marketplace/community-operators-6bjgq" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.215213 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8bfv\" (UniqueName: \"kubernetes.io/projected/a97ed1ff-657f-4bde-943b-78caf9d07f92-kube-api-access-v8bfv\") pod \"certified-operators-74twj\" (UID: \"a97ed1ff-657f-4bde-943b-78caf9d07f92\") " pod="openshift-marketplace/certified-operators-74twj" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.215252 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a97ed1ff-657f-4bde-943b-78caf9d07f92-catalog-content\") pod \"certified-operators-74twj\" (UID: \"a97ed1ff-657f-4bde-943b-78caf9d07f92\") " pod="openshift-marketplace/certified-operators-74twj" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.215277 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a97ed1ff-657f-4bde-943b-78caf9d07f92-utilities\") pod \"certified-operators-74twj\" (UID: \"a97ed1ff-657f-4bde-943b-78caf9d07f92\") " pod="openshift-marketplace/certified-operators-74twj" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.215298 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98a35d03-ef3b-4341-9866-56d12a28aee3-catalog-content\") pod \"community-operators-6bjgq\" (UID: \"98a35d03-ef3b-4341-9866-56d12a28aee3\") " pod="openshift-marketplace/community-operators-6bjgq" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.215328 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 16:12:27 crc kubenswrapper[4714]: E0129 16:12:27.216116 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:27.716102785 +0000 UTC m=+154.236603905 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.216563 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.216850 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a97ed1ff-657f-4bde-943b-78caf9d07f92-catalog-content\") pod \"certified-operators-74twj\" (UID: \"a97ed1ff-657f-4bde-943b-78caf9d07f92\") " pod="openshift-marketplace/certified-operators-74twj" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.216918 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a97ed1ff-657f-4bde-943b-78caf9d07f92-utilities\") pod \"certified-operators-74twj\" (UID: \"a97ed1ff-657f-4bde-943b-78caf9d07f92\") " pod="openshift-marketplace/certified-operators-74twj" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.264895 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.294916 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8bfv\" (UniqueName: \"kubernetes.io/projected/a97ed1ff-657f-4bde-943b-78caf9d07f92-kube-api-access-v8bfv\") pod \"certified-operators-74twj\" (UID: \"a97ed1ff-657f-4bde-943b-78caf9d07f92\") " pod="openshift-marketplace/certified-operators-74twj" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.316443 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.316806 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gks4z\" (UniqueName: \"kubernetes.io/projected/98a35d03-ef3b-4341-9866-56d12a28aee3-kube-api-access-gks4z\") pod \"community-operators-6bjgq\" (UID: \"98a35d03-ef3b-4341-9866-56d12a28aee3\") " pod="openshift-marketplace/community-operators-6bjgq" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.316911 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98a35d03-ef3b-4341-9866-56d12a28aee3-catalog-content\") pod \"community-operators-6bjgq\" (UID: \"98a35d03-ef3b-4341-9866-56d12a28aee3\") " pod="openshift-marketplace/community-operators-6bjgq" Jan 29 16:12:27 crc kubenswrapper[4714]: E0129 16:12:27.317101 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:27.817062102 +0000 UTC m=+154.337563222 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.317245 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.317363 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98a35d03-ef3b-4341-9866-56d12a28aee3-utilities\") pod \"community-operators-6bjgq\" (UID: \"98a35d03-ef3b-4341-9866-56d12a28aee3\") " pod="openshift-marketplace/community-operators-6bjgq" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.318876 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98a35d03-ef3b-4341-9866-56d12a28aee3-catalog-content\") pod \"community-operators-6bjgq\" (UID: \"98a35d03-ef3b-4341-9866-56d12a28aee3\") " pod="openshift-marketplace/community-operators-6bjgq" Jan 29 16:12:27 crc kubenswrapper[4714]: E0129 16:12:27.319323 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:27.819304071 +0000 UTC m=+154.339805281 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.319518 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98a35d03-ef3b-4341-9866-56d12a28aee3-utilities\") pod \"community-operators-6bjgq\" (UID: \"98a35d03-ef3b-4341-9866-56d12a28aee3\") " pod="openshift-marketplace/community-operators-6bjgq" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.339421 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-74twj" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.364249 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dbvgp"] Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.364611 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gks4z\" (UniqueName: \"kubernetes.io/projected/98a35d03-ef3b-4341-9866-56d12a28aee3-kube-api-access-gks4z\") pod \"community-operators-6bjgq\" (UID: \"98a35d03-ef3b-4341-9866-56d12a28aee3\") " pod="openshift-marketplace/community-operators-6bjgq" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.365450 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbvgp" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.377193 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.401547 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dbvgp"] Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.421048 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.421214 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef-utilities\") pod \"certified-operators-dbvgp\" (UID: \"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef\") " pod="openshift-marketplace/certified-operators-dbvgp" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.421234 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef-catalog-content\") pod \"certified-operators-dbvgp\" (UID: \"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef\") " pod="openshift-marketplace/certified-operators-dbvgp" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.421318 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dv2v\" (UniqueName: \"kubernetes.io/projected/ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef-kube-api-access-9dv2v\") pod \"certified-operators-dbvgp\" (UID: \"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef\") " pod="openshift-marketplace/certified-operators-dbvgp" Jan 29 16:12:27 crc kubenswrapper[4714]: E0129 16:12:27.421443 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:27.921428523 +0000 UTC m=+154.441929643 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.461054 4714 generic.go:334] "Generic (PLEG): container finished" podID="fee5819c-8349-4080-9922-453f31a300da" containerID="35db7973ed160da110c671870acde50bc58a4cf360b02511f01c25b5e19a9de9" exitCode=0 Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.461513 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"fee5819c-8349-4080-9922-453f31a300da","Type":"ContainerDied","Data":"35db7973ed160da110c671870acde50bc58a4cf360b02511f01c25b5e19a9de9"} Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.522438 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dv2v\" (UniqueName: \"kubernetes.io/projected/ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef-kube-api-access-9dv2v\") pod \"certified-operators-dbvgp\" (UID: \"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef\") " pod="openshift-marketplace/certified-operators-dbvgp" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.522554 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef-utilities\") pod \"certified-operators-dbvgp\" (UID: \"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef\") " pod="openshift-marketplace/certified-operators-dbvgp" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.522578 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef-catalog-content\") pod \"certified-operators-dbvgp\" (UID: \"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef\") " pod="openshift-marketplace/certified-operators-dbvgp" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.522623 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:27 crc kubenswrapper[4714]: E0129 16:12:27.523575 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:28.023560066 +0000 UTC m=+154.544061186 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.523589 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef-catalog-content\") pod \"certified-operators-dbvgp\" (UID: \"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef\") " pod="openshift-marketplace/certified-operators-dbvgp" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.523980 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef-utilities\") pod \"certified-operators-dbvgp\" (UID: \"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef\") " pod="openshift-marketplace/certified-operators-dbvgp" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.526903 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6bjgq" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.550445 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dv2v\" (UniqueName: \"kubernetes.io/projected/ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef-kube-api-access-9dv2v\") pod \"certified-operators-dbvgp\" (UID: \"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef\") " pod="openshift-marketplace/certified-operators-dbvgp" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.623432 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:27 crc kubenswrapper[4714]: E0129 16:12:27.623855 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:28.1237673 +0000 UTC m=+154.644268420 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.700226 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbvgp" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.710063 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xtr82"] Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.724408 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:27 crc kubenswrapper[4714]: E0129 16:12:27.724797 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:28.224785489 +0000 UTC m=+154.745286609 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.748109 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 16:12:27 crc kubenswrapper[4714]: [-]has-synced failed: reason withheld Jan 29 16:12:27 crc kubenswrapper[4714]: [+]process-running ok Jan 29 16:12:27 crc kubenswrapper[4714]: healthz check failed Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.748154 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 16:12:27 crc kubenswrapper[4714]: W0129 16:12:27.777486 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11a30de8_b234_47b4_8fd0_44f0c428be78.slice/crio-d1e11cf94d1ae7d280d25746da20bca8871b9f9c8323efe87d1cfb324504d7a1 WatchSource:0}: Error finding container d1e11cf94d1ae7d280d25746da20bca8871b9f9c8323efe87d1cfb324504d7a1: Status 404 returned error can't find the container with id d1e11cf94d1ae7d280d25746da20bca8871b9f9c8323efe87d1cfb324504d7a1 Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.791672 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dwsm5" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.827493 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:27 crc kubenswrapper[4714]: E0129 16:12:27.828106 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:28.328078748 +0000 UTC m=+154.848579868 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.853350 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.853394 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.876203 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-74twj"] Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.901193 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.928752 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb979c55-3027-4d92-94b9-cd17c32e6331-config-volume\") pod \"cb979c55-3027-4d92-94b9-cd17c32e6331\" (UID: \"cb979c55-3027-4d92-94b9-cd17c32e6331\") " Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.928832 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fzjx\" (UniqueName: \"kubernetes.io/projected/cb979c55-3027-4d92-94b9-cd17c32e6331-kube-api-access-7fzjx\") pod \"cb979c55-3027-4d92-94b9-cd17c32e6331\" (UID: \"cb979c55-3027-4d92-94b9-cd17c32e6331\") " Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.928969 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb979c55-3027-4d92-94b9-cd17c32e6331-secret-volume\") pod \"cb979c55-3027-4d92-94b9-cd17c32e6331\" (UID: \"cb979c55-3027-4d92-94b9-cd17c32e6331\") " Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.929112 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:27 crc kubenswrapper[4714]: E0129 16:12:27.929437 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:28.429426136 +0000 UTC m=+154.949927256 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.929597 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb979c55-3027-4d92-94b9-cd17c32e6331-config-volume" (OuterVolumeSpecName: "config-volume") pod "cb979c55-3027-4d92-94b9-cd17c32e6331" (UID: "cb979c55-3027-4d92-94b9-cd17c32e6331"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.940080 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb979c55-3027-4d92-94b9-cd17c32e6331-kube-api-access-7fzjx" (OuterVolumeSpecName: "kube-api-access-7fzjx") pod "cb979c55-3027-4d92-94b9-cd17c32e6331" (UID: "cb979c55-3027-4d92-94b9-cd17c32e6331"). InnerVolumeSpecName "kube-api-access-7fzjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:12:27 crc kubenswrapper[4714]: I0129 16:12:27.946248 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb979c55-3027-4d92-94b9-cd17c32e6331-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cb979c55-3027-4d92-94b9-cd17c32e6331" (UID: "cb979c55-3027-4d92-94b9-cd17c32e6331"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.030219 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.030786 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:28 crc kubenswrapper[4714]: E0129 16:12:28.030980 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:28.53093029 +0000 UTC m=+155.051431420 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.031152 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.031253 4714 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb979c55-3027-4d92-94b9-cd17c32e6331-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.031265 4714 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb979c55-3027-4d92-94b9-cd17c32e6331-config-volume\") on node \"crc\" DevicePath \"\"" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.031276 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fzjx\" (UniqueName: \"kubernetes.io/projected/cb979c55-3027-4d92-94b9-cd17c32e6331-kube-api-access-7fzjx\") on node \"crc\" DevicePath \"\"" Jan 29 16:12:28 crc kubenswrapper[4714]: E0129 16:12:28.031563 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:28.531547649 +0000 UTC m=+155.052048769 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.138398 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:28 crc kubenswrapper[4714]: E0129 16:12:28.138729 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:28.638715396 +0000 UTC m=+155.159216516 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.240038 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:28 crc kubenswrapper[4714]: E0129 16:12:28.240338 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:28.740326923 +0000 UTC m=+155.260828033 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.276186 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dbvgp"] Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.338536 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6bjgq"] Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.341062 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:28 crc kubenswrapper[4714]: E0129 16:12:28.341430 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:28.841414444 +0000 UTC m=+155.361915564 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.442295 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:28 crc kubenswrapper[4714]: E0129 16:12:28.442695 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:28.94267872 +0000 UTC m=+155.463179840 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.466345 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bjgq" event={"ID":"98a35d03-ef3b-4341-9866-56d12a28aee3","Type":"ContainerStarted","Data":"d38f58d434dcb4497833c894d88b3ceb4be10c7a4d69f2a5403bda7aa069a88c"} Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.472144 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-74twj" event={"ID":"a97ed1ff-657f-4bde-943b-78caf9d07f92","Type":"ContainerStarted","Data":"27e514e7925336355503e562c2b866089bbb8f20f6235853c55635bfeebcfe8c"} Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.473473 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbvgp" event={"ID":"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef","Type":"ContainerStarted","Data":"8245ba1ef35303ce5087b4ec9f0268e726a450c3c8f0f72042d1655209fffe8b"} Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.474926 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xtr82" event={"ID":"11a30de8-b234-47b4-8fd0-44f0c428be78","Type":"ContainerStarted","Data":"d1e11cf94d1ae7d280d25746da20bca8871b9f9c8323efe87d1cfb324504d7a1"} Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.475899 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58","Type":"ContainerStarted","Data":"787dde46e1e979113da67d8494483beb2242c954a937435fdd6e95f1ea828572"} Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.477700 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.478196 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-5mkf8" event={"ID":"cb979c55-3027-4d92-94b9-cd17c32e6331","Type":"ContainerDied","Data":"64dd0aa019d2ff78b021ae7336dcb6b797f2fc222a74d3ec8b1b6722bf60513e"} Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.478214 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64dd0aa019d2ff78b021ae7336dcb6b797f2fc222a74d3ec8b1b6722bf60513e" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.542990 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:28 crc kubenswrapper[4714]: E0129 16:12:28.543227 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:29.043200794 +0000 UTC m=+155.563701914 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.543280 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:28 crc kubenswrapper[4714]: E0129 16:12:28.543582 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:29.043569315 +0000 UTC m=+155.564070435 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.644889 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:28 crc kubenswrapper[4714]: E0129 16:12:28.645079 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:29.145054518 +0000 UTC m=+155.665555638 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.645115 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:28 crc kubenswrapper[4714]: E0129 16:12:28.645441 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:29.14542858 +0000 UTC m=+155.665929690 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.745770 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:28 crc kubenswrapper[4714]: E0129 16:12:28.745993 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:29.245964724 +0000 UTC m=+155.766465844 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.746114 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:28 crc kubenswrapper[4714]: E0129 16:12:28.746382 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:29.246371406 +0000 UTC m=+155.766872526 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.746781 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nssrv"] Jan 29 16:12:28 crc kubenswrapper[4714]: E0129 16:12:28.746996 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb979c55-3027-4d92-94b9-cd17c32e6331" containerName="collect-profiles" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.747009 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb979c55-3027-4d92-94b9-cd17c32e6331" containerName="collect-profiles" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.747103 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb979c55-3027-4d92-94b9-cd17c32e6331" containerName="collect-profiles" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.747749 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nssrv" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.749815 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.759635 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 16:12:28 crc kubenswrapper[4714]: [-]has-synced failed: reason withheld Jan 29 16:12:28 crc kubenswrapper[4714]: [+]process-running ok Jan 29 16:12:28 crc kubenswrapper[4714]: healthz check failed Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.759674 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.766669 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nssrv"] Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.817513 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.847624 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:28 crc kubenswrapper[4714]: E0129 16:12:28.847843 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:29.347811008 +0000 UTC m=+155.868312138 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.848042 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.848156 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae853ba-61c9-439b-9dc9-21567075f18a-utilities\") pod \"redhat-marketplace-nssrv\" (UID: \"eae853ba-61c9-439b-9dc9-21567075f18a\") " pod="openshift-marketplace/redhat-marketplace-nssrv" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.848193 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae853ba-61c9-439b-9dc9-21567075f18a-catalog-content\") pod \"redhat-marketplace-nssrv\" (UID: \"eae853ba-61c9-439b-9dc9-21567075f18a\") " pod="openshift-marketplace/redhat-marketplace-nssrv" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.848269 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntklh\" (UniqueName: \"kubernetes.io/projected/eae853ba-61c9-439b-9dc9-21567075f18a-kube-api-access-ntklh\") pod \"redhat-marketplace-nssrv\" (UID: \"eae853ba-61c9-439b-9dc9-21567075f18a\") " pod="openshift-marketplace/redhat-marketplace-nssrv" Jan 29 16:12:28 crc kubenswrapper[4714]: E0129 16:12:28.848568 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:29.348557171 +0000 UTC m=+155.869058291 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.949192 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fee5819c-8349-4080-9922-453f31a300da-kubelet-dir\") pod \"fee5819c-8349-4080-9922-453f31a300da\" (UID: \"fee5819c-8349-4080-9922-453f31a300da\") " Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.949333 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fee5819c-8349-4080-9922-453f31a300da-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "fee5819c-8349-4080-9922-453f31a300da" (UID: "fee5819c-8349-4080-9922-453f31a300da"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.949374 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.949416 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fee5819c-8349-4080-9922-453f31a300da-kube-api-access\") pod \"fee5819c-8349-4080-9922-453f31a300da\" (UID: \"fee5819c-8349-4080-9922-453f31a300da\") " Jan 29 16:12:28 crc kubenswrapper[4714]: E0129 16:12:28.949554 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:29.449521108 +0000 UTC m=+155.970022228 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.949632 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.949694 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae853ba-61c9-439b-9dc9-21567075f18a-utilities\") pod \"redhat-marketplace-nssrv\" (UID: \"eae853ba-61c9-439b-9dc9-21567075f18a\") " pod="openshift-marketplace/redhat-marketplace-nssrv" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.949730 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae853ba-61c9-439b-9dc9-21567075f18a-catalog-content\") pod \"redhat-marketplace-nssrv\" (UID: \"eae853ba-61c9-439b-9dc9-21567075f18a\") " pod="openshift-marketplace/redhat-marketplace-nssrv" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.949779 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntklh\" (UniqueName: \"kubernetes.io/projected/eae853ba-61c9-439b-9dc9-21567075f18a-kube-api-access-ntklh\") pod \"redhat-marketplace-nssrv\" (UID: \"eae853ba-61c9-439b-9dc9-21567075f18a\") " pod="openshift-marketplace/redhat-marketplace-nssrv" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.949839 4714 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fee5819c-8349-4080-9922-453f31a300da-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 29 16:12:28 crc kubenswrapper[4714]: E0129 16:12:28.949991 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:29.449982512 +0000 UTC m=+155.970483632 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.950588 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae853ba-61c9-439b-9dc9-21567075f18a-catalog-content\") pod \"redhat-marketplace-nssrv\" (UID: \"eae853ba-61c9-439b-9dc9-21567075f18a\") " pod="openshift-marketplace/redhat-marketplace-nssrv" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.950811 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae853ba-61c9-439b-9dc9-21567075f18a-utilities\") pod \"redhat-marketplace-nssrv\" (UID: \"eae853ba-61c9-439b-9dc9-21567075f18a\") " pod="openshift-marketplace/redhat-marketplace-nssrv" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.954877 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fee5819c-8349-4080-9922-453f31a300da-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "fee5819c-8349-4080-9922-453f31a300da" (UID: "fee5819c-8349-4080-9922-453f31a300da"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:12:28 crc kubenswrapper[4714]: I0129 16:12:28.966415 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntklh\" (UniqueName: \"kubernetes.io/projected/eae853ba-61c9-439b-9dc9-21567075f18a-kube-api-access-ntklh\") pod \"redhat-marketplace-nssrv\" (UID: \"eae853ba-61c9-439b-9dc9-21567075f18a\") " pod="openshift-marketplace/redhat-marketplace-nssrv" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.051092 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:29 crc kubenswrapper[4714]: E0129 16:12:29.051242 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:29.551222008 +0000 UTC m=+156.071723138 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.051425 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:29 crc kubenswrapper[4714]: E0129 16:12:29.051768 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:29.551758174 +0000 UTC m=+156.072259304 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.051966 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fee5819c-8349-4080-9922-453f31a300da-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.062891 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nssrv" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.136294 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hbcnj"] Jan 29 16:12:29 crc kubenswrapper[4714]: E0129 16:12:29.136519 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fee5819c-8349-4080-9922-453f31a300da" containerName="pruner" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.136530 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="fee5819c-8349-4080-9922-453f31a300da" containerName="pruner" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.136639 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="fee5819c-8349-4080-9922-453f31a300da" containerName="pruner" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.137475 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hbcnj" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.152868 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:29 crc kubenswrapper[4714]: E0129 16:12:29.153163 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:29.653136834 +0000 UTC m=+156.173637964 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.153562 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.153635 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/213a402c-b327-4aa6-9690-22d6da8664a4-utilities\") pod \"redhat-marketplace-hbcnj\" (UID: \"213a402c-b327-4aa6-9690-22d6da8664a4\") " pod="openshift-marketplace/redhat-marketplace-hbcnj" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.153682 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pkm4\" (UniqueName: \"kubernetes.io/projected/213a402c-b327-4aa6-9690-22d6da8664a4-kube-api-access-8pkm4\") pod \"redhat-marketplace-hbcnj\" (UID: \"213a402c-b327-4aa6-9690-22d6da8664a4\") " pod="openshift-marketplace/redhat-marketplace-hbcnj" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.153708 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/213a402c-b327-4aa6-9690-22d6da8664a4-catalog-content\") pod \"redhat-marketplace-hbcnj\" (UID: \"213a402c-b327-4aa6-9690-22d6da8664a4\") " pod="openshift-marketplace/redhat-marketplace-hbcnj" Jan 29 16:12:29 crc kubenswrapper[4714]: E0129 16:12:29.154129 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:29.654107144 +0000 UTC m=+156.174608264 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.169917 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hbcnj"] Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.255693 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:29 crc kubenswrapper[4714]: E0129 16:12:29.255874 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:29.755843424 +0000 UTC m=+156.276344544 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.255991 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/213a402c-b327-4aa6-9690-22d6da8664a4-utilities\") pod \"redhat-marketplace-hbcnj\" (UID: \"213a402c-b327-4aa6-9690-22d6da8664a4\") " pod="openshift-marketplace/redhat-marketplace-hbcnj" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.256044 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pkm4\" (UniqueName: \"kubernetes.io/projected/213a402c-b327-4aa6-9690-22d6da8664a4-kube-api-access-8pkm4\") pod \"redhat-marketplace-hbcnj\" (UID: \"213a402c-b327-4aa6-9690-22d6da8664a4\") " pod="openshift-marketplace/redhat-marketplace-hbcnj" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.256072 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/213a402c-b327-4aa6-9690-22d6da8664a4-catalog-content\") pod \"redhat-marketplace-hbcnj\" (UID: \"213a402c-b327-4aa6-9690-22d6da8664a4\") " pod="openshift-marketplace/redhat-marketplace-hbcnj" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.256125 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:29 crc kubenswrapper[4714]: E0129 16:12:29.256541 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:29.756525125 +0000 UTC m=+156.277026245 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.256559 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/213a402c-b327-4aa6-9690-22d6da8664a4-utilities\") pod \"redhat-marketplace-hbcnj\" (UID: \"213a402c-b327-4aa6-9690-22d6da8664a4\") " pod="openshift-marketplace/redhat-marketplace-hbcnj" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.256576 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/213a402c-b327-4aa6-9690-22d6da8664a4-catalog-content\") pod \"redhat-marketplace-hbcnj\" (UID: \"213a402c-b327-4aa6-9690-22d6da8664a4\") " pod="openshift-marketplace/redhat-marketplace-hbcnj" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.273953 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pkm4\" (UniqueName: \"kubernetes.io/projected/213a402c-b327-4aa6-9690-22d6da8664a4-kube-api-access-8pkm4\") pod \"redhat-marketplace-hbcnj\" (UID: \"213a402c-b327-4aa6-9690-22d6da8664a4\") " pod="openshift-marketplace/redhat-marketplace-hbcnj" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.290595 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nssrv"] Jan 29 16:12:29 crc kubenswrapper[4714]: W0129 16:12:29.303756 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeae853ba_61c9_439b_9dc9_21567075f18a.slice/crio-11eca2d99e975c8d4c6d498c418a6ed86174580092ad733d4cf31d057f9d974e WatchSource:0}: Error finding container 11eca2d99e975c8d4c6d498c418a6ed86174580092ad733d4cf31d057f9d974e: Status 404 returned error can't find the container with id 11eca2d99e975c8d4c6d498c418a6ed86174580092ad733d4cf31d057f9d974e Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.357989 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:29 crc kubenswrapper[4714]: E0129 16:12:29.358196 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:29.858171403 +0000 UTC m=+156.378672523 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.358421 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:29 crc kubenswrapper[4714]: E0129 16:12:29.358757 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:29.858742981 +0000 UTC m=+156.379244101 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.459037 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:29 crc kubenswrapper[4714]: E0129 16:12:29.459184 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:29.959167091 +0000 UTC m=+156.479668211 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.459331 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:29 crc kubenswrapper[4714]: E0129 16:12:29.459580 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:29.959572134 +0000 UTC m=+156.480073254 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.463000 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hbcnj" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.483384 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nssrv" event={"ID":"eae853ba-61c9-439b-9dc9-21567075f18a","Type":"ContainerStarted","Data":"11eca2d99e975c8d4c6d498c418a6ed86174580092ad733d4cf31d057f9d974e"} Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.484875 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"fee5819c-8349-4080-9922-453f31a300da","Type":"ContainerDied","Data":"aabce909ed74e646558bee9fa02988139c96b4993b0d41ddfcd7351dba0ff624"} Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.484922 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aabce909ed74e646558bee9fa02988139c96b4993b0d41ddfcd7351dba0ff624" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.484969 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.559977 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:29 crc kubenswrapper[4714]: E0129 16:12:29.560184 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:30.060152699 +0000 UTC m=+156.580653829 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.560605 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:29 crc kubenswrapper[4714]: E0129 16:12:29.560978 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:30.060967304 +0000 UTC m=+156.581468444 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.661500 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:29 crc kubenswrapper[4714]: E0129 16:12:29.661679 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:30.161651533 +0000 UTC m=+156.682152663 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.661867 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:29 crc kubenswrapper[4714]: E0129 16:12:29.662233 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:30.16221971 +0000 UTC m=+156.682720830 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.671079 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hbcnj"] Jan 29 16:12:29 crc kubenswrapper[4714]: W0129 16:12:29.679455 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod213a402c_b327_4aa6_9690_22d6da8664a4.slice/crio-cc206117744230d0505236f4ba4b88035d78daaa5da318d54c519b3dd8b10d4e WatchSource:0}: Error finding container cc206117744230d0505236f4ba4b88035d78daaa5da318d54c519b3dd8b10d4e: Status 404 returned error can't find the container with id cc206117744230d0505236f4ba4b88035d78daaa5da318d54c519b3dd8b10d4e Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.743677 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 16:12:29 crc kubenswrapper[4714]: [-]has-synced failed: reason withheld Jan 29 16:12:29 crc kubenswrapper[4714]: [+]process-running ok Jan 29 16:12:29 crc kubenswrapper[4714]: healthz check failed Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.743744 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.762748 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:29 crc kubenswrapper[4714]: E0129 16:12:29.762915 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:30.262894539 +0000 UTC m=+156.783395669 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.763272 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:29 crc kubenswrapper[4714]: E0129 16:12:29.763640 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:30.263628461 +0000 UTC m=+156.784129591 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.864769 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:29 crc kubenswrapper[4714]: E0129 16:12:29.864915 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:30.364892736 +0000 UTC m=+156.885393856 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.865258 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:29 crc kubenswrapper[4714]: E0129 16:12:29.865560 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:30.365553007 +0000 UTC m=+156.886054127 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.938102 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lb68h"] Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.945379 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lb68h" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.951193 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.963899 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lb68h"] Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.973970 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.974230 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d05e7c79-7d66-4453-aedb-f240784ff294-utilities\") pod \"redhat-operators-lb68h\" (UID: \"d05e7c79-7d66-4453-aedb-f240784ff294\") " pod="openshift-marketplace/redhat-operators-lb68h" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.974276 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4jjj\" (UniqueName: \"kubernetes.io/projected/d05e7c79-7d66-4453-aedb-f240784ff294-kube-api-access-m4jjj\") pod \"redhat-operators-lb68h\" (UID: \"d05e7c79-7d66-4453-aedb-f240784ff294\") " pod="openshift-marketplace/redhat-operators-lb68h" Jan 29 16:12:29 crc kubenswrapper[4714]: I0129 16:12:29.974297 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d05e7c79-7d66-4453-aedb-f240784ff294-catalog-content\") pod \"redhat-operators-lb68h\" (UID: \"d05e7c79-7d66-4453-aedb-f240784ff294\") " pod="openshift-marketplace/redhat-operators-lb68h" Jan 29 16:12:29 crc kubenswrapper[4714]: E0129 16:12:29.974439 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:30.474424066 +0000 UTC m=+156.994925186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.076512 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.076603 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d05e7c79-7d66-4453-aedb-f240784ff294-utilities\") pod \"redhat-operators-lb68h\" (UID: \"d05e7c79-7d66-4453-aedb-f240784ff294\") " pod="openshift-marketplace/redhat-operators-lb68h" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.076692 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4jjj\" (UniqueName: \"kubernetes.io/projected/d05e7c79-7d66-4453-aedb-f240784ff294-kube-api-access-m4jjj\") pod \"redhat-operators-lb68h\" (UID: \"d05e7c79-7d66-4453-aedb-f240784ff294\") " pod="openshift-marketplace/redhat-operators-lb68h" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.076730 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d05e7c79-7d66-4453-aedb-f240784ff294-catalog-content\") pod \"redhat-operators-lb68h\" (UID: \"d05e7c79-7d66-4453-aedb-f240784ff294\") " pod="openshift-marketplace/redhat-operators-lb68h" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.077283 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d05e7c79-7d66-4453-aedb-f240784ff294-catalog-content\") pod \"redhat-operators-lb68h\" (UID: \"d05e7c79-7d66-4453-aedb-f240784ff294\") " pod="openshift-marketplace/redhat-operators-lb68h" Jan 29 16:12:30 crc kubenswrapper[4714]: E0129 16:12:30.077368 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:30.577348913 +0000 UTC m=+157.097850043 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.077639 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d05e7c79-7d66-4453-aedb-f240784ff294-utilities\") pod \"redhat-operators-lb68h\" (UID: \"d05e7c79-7d66-4453-aedb-f240784ff294\") " pod="openshift-marketplace/redhat-operators-lb68h" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.096443 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4jjj\" (UniqueName: \"kubernetes.io/projected/d05e7c79-7d66-4453-aedb-f240784ff294-kube-api-access-m4jjj\") pod \"redhat-operators-lb68h\" (UID: \"d05e7c79-7d66-4453-aedb-f240784ff294\") " pod="openshift-marketplace/redhat-operators-lb68h" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.177722 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:30 crc kubenswrapper[4714]: E0129 16:12:30.177898 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:30.677870236 +0000 UTC m=+157.198371366 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.178189 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:30 crc kubenswrapper[4714]: E0129 16:12:30.178582 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:30.678567398 +0000 UTC m=+157.199068518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.262510 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lb68h" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.279641 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:30 crc kubenswrapper[4714]: E0129 16:12:30.279833 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:30.779794223 +0000 UTC m=+157.300295353 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.280281 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:30 crc kubenswrapper[4714]: E0129 16:12:30.280681 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:30.78067068 +0000 UTC m=+157.301171810 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.349067 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kg9qt"] Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.359220 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kg9qt" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.362016 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kg9qt"] Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.382587 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:30 crc kubenswrapper[4714]: E0129 16:12:30.382836 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:30.882808593 +0000 UTC m=+157.403309723 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.383207 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpzw5\" (UniqueName: \"kubernetes.io/projected/ec0eba7e-2ea0-432f-bc57-d87404801abe-kube-api-access-xpzw5\") pod \"redhat-operators-kg9qt\" (UID: \"ec0eba7e-2ea0-432f-bc57-d87404801abe\") " pod="openshift-marketplace/redhat-operators-kg9qt" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.383254 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec0eba7e-2ea0-432f-bc57-d87404801abe-utilities\") pod \"redhat-operators-kg9qt\" (UID: \"ec0eba7e-2ea0-432f-bc57-d87404801abe\") " pod="openshift-marketplace/redhat-operators-kg9qt" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.383294 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.383326 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec0eba7e-2ea0-432f-bc57-d87404801abe-catalog-content\") pod \"redhat-operators-kg9qt\" (UID: \"ec0eba7e-2ea0-432f-bc57-d87404801abe\") " pod="openshift-marketplace/redhat-operators-kg9qt" Jan 29 16:12:30 crc kubenswrapper[4714]: E0129 16:12:30.383662 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:30.883648118 +0000 UTC m=+157.404149238 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.478970 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lb68h"] Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.484550 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:30 crc kubenswrapper[4714]: E0129 16:12:30.484731 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:30.984685718 +0000 UTC m=+157.505186838 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.484806 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec0eba7e-2ea0-432f-bc57-d87404801abe-catalog-content\") pod \"redhat-operators-kg9qt\" (UID: \"ec0eba7e-2ea0-432f-bc57-d87404801abe\") " pod="openshift-marketplace/redhat-operators-kg9qt" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.484928 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpzw5\" (UniqueName: \"kubernetes.io/projected/ec0eba7e-2ea0-432f-bc57-d87404801abe-kube-api-access-xpzw5\") pod \"redhat-operators-kg9qt\" (UID: \"ec0eba7e-2ea0-432f-bc57-d87404801abe\") " pod="openshift-marketplace/redhat-operators-kg9qt" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.484979 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec0eba7e-2ea0-432f-bc57-d87404801abe-utilities\") pod \"redhat-operators-kg9qt\" (UID: \"ec0eba7e-2ea0-432f-bc57-d87404801abe\") " pod="openshift-marketplace/redhat-operators-kg9qt" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.485018 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:30 crc kubenswrapper[4714]: E0129 16:12:30.485335 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:30.985320927 +0000 UTC m=+157.505822047 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.485846 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec0eba7e-2ea0-432f-bc57-d87404801abe-catalog-content\") pod \"redhat-operators-kg9qt\" (UID: \"ec0eba7e-2ea0-432f-bc57-d87404801abe\") " pod="openshift-marketplace/redhat-operators-kg9qt" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.486443 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec0eba7e-2ea0-432f-bc57-d87404801abe-utilities\") pod \"redhat-operators-kg9qt\" (UID: \"ec0eba7e-2ea0-432f-bc57-d87404801abe\") " pod="openshift-marketplace/redhat-operators-kg9qt" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.500722 4714 generic.go:334] "Generic (PLEG): container finished" podID="213a402c-b327-4aa6-9690-22d6da8664a4" containerID="53e99ef17ef7a89695643694f24aab5f9e6445925a8b86fd6d021c08cb6c082f" exitCode=0 Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.500842 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hbcnj" event={"ID":"213a402c-b327-4aa6-9690-22d6da8664a4","Type":"ContainerDied","Data":"53e99ef17ef7a89695643694f24aab5f9e6445925a8b86fd6d021c08cb6c082f"} Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.501054 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hbcnj" event={"ID":"213a402c-b327-4aa6-9690-22d6da8664a4","Type":"ContainerStarted","Data":"cc206117744230d0505236f4ba4b88035d78daaa5da318d54c519b3dd8b10d4e"} Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.502674 4714 generic.go:334] "Generic (PLEG): container finished" podID="98a35d03-ef3b-4341-9866-56d12a28aee3" containerID="7d5da21ddc846f7484a829e4e1b7d4f27ddad6196e5ccbce162fd0a2651869dd" exitCode=0 Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.502751 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bjgq" event={"ID":"98a35d03-ef3b-4341-9866-56d12a28aee3","Type":"ContainerDied","Data":"7d5da21ddc846f7484a829e4e1b7d4f27ddad6196e5ccbce162fd0a2651869dd"} Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.505235 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpzw5\" (UniqueName: \"kubernetes.io/projected/ec0eba7e-2ea0-432f-bc57-d87404801abe-kube-api-access-xpzw5\") pod \"redhat-operators-kg9qt\" (UID: \"ec0eba7e-2ea0-432f-bc57-d87404801abe\") " pod="openshift-marketplace/redhat-operators-kg9qt" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.508234 4714 generic.go:334] "Generic (PLEG): container finished" podID="a97ed1ff-657f-4bde-943b-78caf9d07f92" containerID="bebaa193fd909649d996ac5ecb12f75e7aa251dd1a4b1b911882734c87e4b046" exitCode=0 Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.508295 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-74twj" event={"ID":"a97ed1ff-657f-4bde-943b-78caf9d07f92","Type":"ContainerDied","Data":"bebaa193fd909649d996ac5ecb12f75e7aa251dd1a4b1b911882734c87e4b046"} Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.509730 4714 generic.go:334] "Generic (PLEG): container finished" podID="ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef" containerID="bf193fd80bff7bed1bec1edfb59432d0f18ec27217fda44032cd6a47058aee41" exitCode=0 Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.509813 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbvgp" event={"ID":"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef","Type":"ContainerDied","Data":"bf193fd80bff7bed1bec1edfb59432d0f18ec27217fda44032cd6a47058aee41"} Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.510441 4714 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.514471 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lb68h" event={"ID":"d05e7c79-7d66-4453-aedb-f240784ff294","Type":"ContainerStarted","Data":"5e22f2e727671a2879c86dcb9146aebbe76ddedf77fd5e705c834b21cf8bd941"} Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.521966 4714 generic.go:334] "Generic (PLEG): container finished" podID="eae853ba-61c9-439b-9dc9-21567075f18a" containerID="4ddfcc2a03d228c12293d91791425de7e03e1ce7bf286f993aef049f853f58fe" exitCode=0 Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.522109 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nssrv" event={"ID":"eae853ba-61c9-439b-9dc9-21567075f18a","Type":"ContainerDied","Data":"4ddfcc2a03d228c12293d91791425de7e03e1ce7bf286f993aef049f853f58fe"} Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.526192 4714 generic.go:334] "Generic (PLEG): container finished" podID="11a30de8-b234-47b4-8fd0-44f0c428be78" containerID="70b956b273a4676b7f7a5e461f43c09f1067f834a7d754d57cae60e15152821d" exitCode=0 Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.526279 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xtr82" event={"ID":"11a30de8-b234-47b4-8fd0-44f0c428be78","Type":"ContainerDied","Data":"70b956b273a4676b7f7a5e461f43c09f1067f834a7d754d57cae60e15152821d"} Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.531506 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58","Type":"ContainerStarted","Data":"25d6585e3011dd0c35de85182524e231475d07c340c6eadfb866c1349154b360"} Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.578420 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.578398993 podStartE2EDuration="3.578398993s" podCreationTimestamp="2026-01-29 16:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:30.576825395 +0000 UTC m=+157.097326535" watchObservedRunningTime="2026-01-29 16:12:30.578398993 +0000 UTC m=+157.098900113" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.586991 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:30 crc kubenswrapper[4714]: E0129 16:12:30.587190 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:31.087153121 +0000 UTC m=+157.607654251 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.587455 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:30 crc kubenswrapper[4714]: E0129 16:12:30.590925 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:31.090905946 +0000 UTC m=+157.611407076 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:30 crc kubenswrapper[4714]: E0129 16:12:30.688823 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:31.188784839 +0000 UTC m=+157.709285999 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.688615 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.689639 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:30 crc kubenswrapper[4714]: E0129 16:12:30.690601 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:31.190579314 +0000 UTC m=+157.711080474 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.691390 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.699045 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgl5s" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.705378 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kg9qt" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.746623 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 16:12:30 crc kubenswrapper[4714]: [-]has-synced failed: reason withheld Jan 29 16:12:30 crc kubenswrapper[4714]: [+]process-running ok Jan 29 16:12:30 crc kubenswrapper[4714]: healthz check failed Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.746692 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.790961 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:30 crc kubenswrapper[4714]: E0129 16:12:30.792310 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:31.292290174 +0000 UTC m=+157.812791294 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.893408 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:30 crc kubenswrapper[4714]: E0129 16:12:30.893871 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:31.393849629 +0000 UTC m=+157.914350829 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.994812 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:30 crc kubenswrapper[4714]: E0129 16:12:30.995096 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:31.495071954 +0000 UTC m=+158.015573074 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:30 crc kubenswrapper[4714]: I0129 16:12:30.995583 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:30 crc kubenswrapper[4714]: E0129 16:12:30.996002 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:31.495986692 +0000 UTC m=+158.016487812 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.016333 4714 patch_prober.go:28] interesting pod/apiserver-76f77b778f-6jl75 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 29 16:12:31 crc kubenswrapper[4714]: [+]log ok Jan 29 16:12:31 crc kubenswrapper[4714]: [+]etcd ok Jan 29 16:12:31 crc kubenswrapper[4714]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 29 16:12:31 crc kubenswrapper[4714]: [+]poststarthook/generic-apiserver-start-informers ok Jan 29 16:12:31 crc kubenswrapper[4714]: [+]poststarthook/max-in-flight-filter ok Jan 29 16:12:31 crc kubenswrapper[4714]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 29 16:12:31 crc kubenswrapper[4714]: [+]poststarthook/image.openshift.io-apiserver-caches ok Jan 29 16:12:31 crc kubenswrapper[4714]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Jan 29 16:12:31 crc kubenswrapper[4714]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Jan 29 16:12:31 crc kubenswrapper[4714]: [+]poststarthook/project.openshift.io-projectcache ok Jan 29 16:12:31 crc kubenswrapper[4714]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Jan 29 16:12:31 crc kubenswrapper[4714]: [+]poststarthook/openshift.io-startinformers ok Jan 29 16:12:31 crc kubenswrapper[4714]: [+]poststarthook/openshift.io-restmapperupdater ok Jan 29 16:12:31 crc kubenswrapper[4714]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 29 16:12:31 crc kubenswrapper[4714]: livez check failed Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.016400 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-6jl75" podUID="99bab267-639b-48b1-abc4-8c0373200a39" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.019700 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kg9qt"] Jan 29 16:12:31 crc kubenswrapper[4714]: W0129 16:12:31.033494 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec0eba7e_2ea0_432f_bc57_d87404801abe.slice/crio-0b72081696a87dc8bb1d6a9fb5d18f85ed353808c0ac8086b9bf62a3458b3736 WatchSource:0}: Error finding container 0b72081696a87dc8bb1d6a9fb5d18f85ed353808c0ac8086b9bf62a3458b3736: Status 404 returned error can't find the container with id 0b72081696a87dc8bb1d6a9fb5d18f85ed353808c0ac8086b9bf62a3458b3736 Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.097273 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:31 crc kubenswrapper[4714]: E0129 16:12:31.097672 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:31.597653591 +0000 UTC m=+158.118154711 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.200006 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:31 crc kubenswrapper[4714]: E0129 16:12:31.200479 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:31.700459364 +0000 UTC m=+158.220960554 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.301028 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:31 crc kubenswrapper[4714]: E0129 16:12:31.301268 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:31.801233265 +0000 UTC m=+158.321734385 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.301349 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:31 crc kubenswrapper[4714]: E0129 16:12:31.301886 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:31.801874985 +0000 UTC m=+158.322376105 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.403634 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:31 crc kubenswrapper[4714]: E0129 16:12:31.403830 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:31.903801612 +0000 UTC m=+158.424302742 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.403902 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:31 crc kubenswrapper[4714]: E0129 16:12:31.404237 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:31.904227355 +0000 UTC m=+158.424728485 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.505311 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:31 crc kubenswrapper[4714]: E0129 16:12:31.505532 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:32.005486051 +0000 UTC m=+158.525987171 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.505952 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:31 crc kubenswrapper[4714]: E0129 16:12:31.506261 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:32.006253584 +0000 UTC m=+158.526754704 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.537737 4714 generic.go:334] "Generic (PLEG): container finished" podID="01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58" containerID="25d6585e3011dd0c35de85182524e231475d07c340c6eadfb866c1349154b360" exitCode=0 Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.537840 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58","Type":"ContainerDied","Data":"25d6585e3011dd0c35de85182524e231475d07c340c6eadfb866c1349154b360"} Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.539044 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg9qt" event={"ID":"ec0eba7e-2ea0-432f-bc57-d87404801abe","Type":"ContainerStarted","Data":"0b72081696a87dc8bb1d6a9fb5d18f85ed353808c0ac8086b9bf62a3458b3736"} Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.542129 4714 generic.go:334] "Generic (PLEG): container finished" podID="d05e7c79-7d66-4453-aedb-f240784ff294" containerID="0e066cf92f43693eba9898b6ef36d8d3eb21b4fa7c877b99db7e3e39dddda134" exitCode=0 Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.542179 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lb68h" event={"ID":"d05e7c79-7d66-4453-aedb-f240784ff294","Type":"ContainerDied","Data":"0e066cf92f43693eba9898b6ef36d8d3eb21b4fa7c877b99db7e3e39dddda134"} Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.545310 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4nghl" event={"ID":"714cef39-2960-4a25-ac81-a4e65a115eb3","Type":"ContainerStarted","Data":"c06e7561b8f72b2f2fba280eae8d647a4fbaf5eabdf2bc315957b0e8c936e176"} Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.607608 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:31 crc kubenswrapper[4714]: E0129 16:12:31.607781 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:32.107750568 +0000 UTC m=+158.628251688 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.608274 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:31 crc kubenswrapper[4714]: E0129 16:12:31.610009 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:32.109993686 +0000 UTC m=+158.630494806 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.709858 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:31 crc kubenswrapper[4714]: E0129 16:12:31.710004 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:32.209975734 +0000 UTC m=+158.730476864 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.710138 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:31 crc kubenswrapper[4714]: E0129 16:12:31.710435 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 16:12:32.210425937 +0000 UTC m=+158.730927147 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gnjmm" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.729322 4714 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.740198 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 16:12:31 crc kubenswrapper[4714]: [-]has-synced failed: reason withheld Jan 29 16:12:31 crc kubenswrapper[4714]: [+]process-running ok Jan 29 16:12:31 crc kubenswrapper[4714]: healthz check failed Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.740258 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.811463 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:31 crc kubenswrapper[4714]: E0129 16:12:31.812004 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 16:12:32.311985353 +0000 UTC m=+158.832486473 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.899950 4714 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-01-29T16:12:31.729348236Z","Handler":null,"Name":""} Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.902302 4714 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.902331 4714 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.913484 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.918543 4714 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 29 16:12:31 crc kubenswrapper[4714]: I0129 16:12:31.918604 4714 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:32 crc kubenswrapper[4714]: I0129 16:12:32.000537 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gnjmm\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:32 crc kubenswrapper[4714]: I0129 16:12:32.014609 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 16:12:32 crc kubenswrapper[4714]: I0129 16:12:32.021613 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:32 crc kubenswrapper[4714]: I0129 16:12:32.084037 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 29 16:12:32 crc kubenswrapper[4714]: I0129 16:12:32.199338 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Jan 29 16:12:32 crc kubenswrapper[4714]: I0129 16:12:32.200548 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-jcdhl" Jan 29 16:12:32 crc kubenswrapper[4714]: I0129 16:12:32.261130 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gnjmm"] Jan 29 16:12:32 crc kubenswrapper[4714]: I0129 16:12:32.525396 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:12:32 crc kubenswrapper[4714]: I0129 16:12:32.550474 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" event={"ID":"48be8ad8-4c02-4bea-a143-449763b39d54","Type":"ContainerStarted","Data":"815b16152db25222f3f6a5ff40233d8cdbe464e73d20d130a327746193531954"} Jan 29 16:12:32 crc kubenswrapper[4714]: I0129 16:12:32.551679 4714 generic.go:334] "Generic (PLEG): container finished" podID="ec0eba7e-2ea0-432f-bc57-d87404801abe" containerID="52ca09e52af15343909984359236a933ffe1f46ab7ada48929cc2741cd10782f" exitCode=0 Jan 29 16:12:32 crc kubenswrapper[4714]: I0129 16:12:32.551752 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg9qt" event={"ID":"ec0eba7e-2ea0-432f-bc57-d87404801abe","Type":"ContainerDied","Data":"52ca09e52af15343909984359236a933ffe1f46ab7ada48929cc2741cd10782f"} Jan 29 16:12:32 crc kubenswrapper[4714]: I0129 16:12:32.743362 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 16:12:32 crc kubenswrapper[4714]: [-]has-synced failed: reason withheld Jan 29 16:12:32 crc kubenswrapper[4714]: [+]process-running ok Jan 29 16:12:32 crc kubenswrapper[4714]: healthz check failed Jan 29 16:12:32 crc kubenswrapper[4714]: I0129 16:12:32.743666 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 16:12:32 crc kubenswrapper[4714]: I0129 16:12:32.765552 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 16:12:32 crc kubenswrapper[4714]: I0129 16:12:32.825176 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58-kube-api-access\") pod \"01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58\" (UID: \"01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58\") " Jan 29 16:12:32 crc kubenswrapper[4714]: I0129 16:12:32.825221 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58-kubelet-dir\") pod \"01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58\" (UID: \"01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58\") " Jan 29 16:12:32 crc kubenswrapper[4714]: I0129 16:12:32.825505 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58" (UID: "01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:12:32 crc kubenswrapper[4714]: I0129 16:12:32.833203 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58" (UID: "01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:12:32 crc kubenswrapper[4714]: I0129 16:12:32.927407 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 16:12:32 crc kubenswrapper[4714]: I0129 16:12:32.927462 4714 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 29 16:12:33 crc kubenswrapper[4714]: I0129 16:12:33.561286 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58","Type":"ContainerDied","Data":"787dde46e1e979113da67d8494483beb2242c954a937435fdd6e95f1ea828572"} Jan 29 16:12:33 crc kubenswrapper[4714]: I0129 16:12:33.561331 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="787dde46e1e979113da67d8494483beb2242c954a937435fdd6e95f1ea828572" Jan 29 16:12:33 crc kubenswrapper[4714]: I0129 16:12:33.561414 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 16:12:33 crc kubenswrapper[4714]: I0129 16:12:33.566871 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4nghl" event={"ID":"714cef39-2960-4a25-ac81-a4e65a115eb3","Type":"ContainerStarted","Data":"6c7aed45b1136fa062f1742097a1e69e9fdfa62a7f2bceaffd7f34de9a0f9750"} Jan 29 16:12:33 crc kubenswrapper[4714]: I0129 16:12:33.741974 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 16:12:33 crc kubenswrapper[4714]: [-]has-synced failed: reason withheld Jan 29 16:12:33 crc kubenswrapper[4714]: [+]process-running ok Jan 29 16:12:33 crc kubenswrapper[4714]: healthz check failed Jan 29 16:12:33 crc kubenswrapper[4714]: I0129 16:12:33.742200 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 16:12:34 crc kubenswrapper[4714]: I0129 16:12:34.575956 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" event={"ID":"48be8ad8-4c02-4bea-a143-449763b39d54","Type":"ContainerStarted","Data":"f2810d7fe0cd9711b0f9c4dd754cc2bb22760a85d91d92553df7afe7b8378eac"} Jan 29 16:12:34 crc kubenswrapper[4714]: I0129 16:12:34.576456 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:34 crc kubenswrapper[4714]: I0129 16:12:34.604987 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" podStartSLOduration=140.604964143 podStartE2EDuration="2m20.604964143s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:34.601585699 +0000 UTC m=+161.122086829" watchObservedRunningTime="2026-01-29 16:12:34.604964143 +0000 UTC m=+161.125465273" Jan 29 16:12:34 crc kubenswrapper[4714]: I0129 16:12:34.741171 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 16:12:34 crc kubenswrapper[4714]: [-]has-synced failed: reason withheld Jan 29 16:12:34 crc kubenswrapper[4714]: [+]process-running ok Jan 29 16:12:34 crc kubenswrapper[4714]: healthz check failed Jan 29 16:12:34 crc kubenswrapper[4714]: I0129 16:12:34.741522 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 16:12:35 crc kubenswrapper[4714]: I0129 16:12:35.586719 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4nghl" event={"ID":"714cef39-2960-4a25-ac81-a4e65a115eb3","Type":"ContainerStarted","Data":"b95620bf86f7dddda0d975ceb522dc33f20f0909fecf473ff8d60f7f216dd939"} Jan 29 16:12:35 crc kubenswrapper[4714]: I0129 16:12:35.614873 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-4nghl" podStartSLOduration=22.614847912 podStartE2EDuration="22.614847912s" podCreationTimestamp="2026-01-29 16:12:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:12:35.612800819 +0000 UTC m=+162.133301949" watchObservedRunningTime="2026-01-29 16:12:35.614847912 +0000 UTC m=+162.135349042" Jan 29 16:12:35 crc kubenswrapper[4714]: I0129 16:12:35.717381 4714 patch_prober.go:28] interesting pod/console-f9d7485db-m2g9h container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Jan 29 16:12:35 crc kubenswrapper[4714]: I0129 16:12:35.717432 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-m2g9h" podUID="0e2a789d-6a90-4d60-881e-9562cd92e0a7" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Jan 29 16:12:35 crc kubenswrapper[4714]: I0129 16:12:35.740410 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 16:12:35 crc kubenswrapper[4714]: [-]has-synced failed: reason withheld Jan 29 16:12:35 crc kubenswrapper[4714]: [+]process-running ok Jan 29 16:12:35 crc kubenswrapper[4714]: healthz check failed Jan 29 16:12:35 crc kubenswrapper[4714]: I0129 16:12:35.740471 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 16:12:35 crc kubenswrapper[4714]: I0129 16:12:35.796090 4714 patch_prober.go:28] interesting pod/downloads-7954f5f757-fn75b container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 16:12:35 crc kubenswrapper[4714]: I0129 16:12:35.796158 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fn75b" podUID="42b66dc3-a385-4350-a943-50f062da35f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 16:12:35 crc kubenswrapper[4714]: I0129 16:12:35.797128 4714 patch_prober.go:28] interesting pod/downloads-7954f5f757-fn75b container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 16:12:35 crc kubenswrapper[4714]: I0129 16:12:35.797272 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-fn75b" podUID="42b66dc3-a385-4350-a943-50f062da35f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 16:12:35 crc kubenswrapper[4714]: I0129 16:12:35.998062 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:36 crc kubenswrapper[4714]: I0129 16:12:36.003787 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-6jl75" Jan 29 16:12:36 crc kubenswrapper[4714]: I0129 16:12:36.740103 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 16:12:36 crc kubenswrapper[4714]: [-]has-synced failed: reason withheld Jan 29 16:12:36 crc kubenswrapper[4714]: [+]process-running ok Jan 29 16:12:36 crc kubenswrapper[4714]: healthz check failed Jan 29 16:12:36 crc kubenswrapper[4714]: I0129 16:12:36.740479 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 16:12:37 crc kubenswrapper[4714]: I0129 16:12:37.739857 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 16:12:37 crc kubenswrapper[4714]: [-]has-synced failed: reason withheld Jan 29 16:12:37 crc kubenswrapper[4714]: [+]process-running ok Jan 29 16:12:37 crc kubenswrapper[4714]: healthz check failed Jan 29 16:12:37 crc kubenswrapper[4714]: I0129 16:12:37.739952 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 16:12:38 crc kubenswrapper[4714]: I0129 16:12:38.103825 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs\") pod \"network-metrics-daemon-2w92b\" (UID: \"791456e8-8d95-4cdb-8fd1-d06a7586b328\") " pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:12:38 crc kubenswrapper[4714]: I0129 16:12:38.116560 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/791456e8-8d95-4cdb-8fd1-d06a7586b328-metrics-certs\") pod \"network-metrics-daemon-2w92b\" (UID: \"791456e8-8d95-4cdb-8fd1-d06a7586b328\") " pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:12:38 crc kubenswrapper[4714]: I0129 16:12:38.407222 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2w92b" Jan 29 16:12:38 crc kubenswrapper[4714]: I0129 16:12:38.719404 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-2w92b"] Jan 29 16:12:38 crc kubenswrapper[4714]: I0129 16:12:38.740591 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 16:12:38 crc kubenswrapper[4714]: [-]has-synced failed: reason withheld Jan 29 16:12:38 crc kubenswrapper[4714]: [+]process-running ok Jan 29 16:12:38 crc kubenswrapper[4714]: healthz check failed Jan 29 16:12:38 crc kubenswrapper[4714]: I0129 16:12:38.740649 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 16:12:39 crc kubenswrapper[4714]: I0129 16:12:39.622647 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2w92b" event={"ID":"791456e8-8d95-4cdb-8fd1-d06a7586b328","Type":"ContainerStarted","Data":"6555557490d8db02ec80c5595e01191e525ec03f524bc437c5c49a933664722a"} Jan 29 16:12:39 crc kubenswrapper[4714]: I0129 16:12:39.739730 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 16:12:39 crc kubenswrapper[4714]: [-]has-synced failed: reason withheld Jan 29 16:12:39 crc kubenswrapper[4714]: [+]process-running ok Jan 29 16:12:39 crc kubenswrapper[4714]: healthz check failed Jan 29 16:12:39 crc kubenswrapper[4714]: I0129 16:12:39.739780 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 16:12:40 crc kubenswrapper[4714]: I0129 16:12:40.628079 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2w92b" event={"ID":"791456e8-8d95-4cdb-8fd1-d06a7586b328","Type":"ContainerStarted","Data":"5aab667586997dedef5a158b01344ff3e116d1ce4032091162e3572b9f0a1729"} Jan 29 16:12:40 crc kubenswrapper[4714]: I0129 16:12:40.739823 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 16:12:40 crc kubenswrapper[4714]: [-]has-synced failed: reason withheld Jan 29 16:12:40 crc kubenswrapper[4714]: [+]process-running ok Jan 29 16:12:40 crc kubenswrapper[4714]: healthz check failed Jan 29 16:12:40 crc kubenswrapper[4714]: I0129 16:12:40.739893 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 16:12:41 crc kubenswrapper[4714]: I0129 16:12:41.739502 4714 patch_prober.go:28] interesting pod/router-default-5444994796-lz6mw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 16:12:41 crc kubenswrapper[4714]: [-]has-synced failed: reason withheld Jan 29 16:12:41 crc kubenswrapper[4714]: [+]process-running ok Jan 29 16:12:41 crc kubenswrapper[4714]: healthz check failed Jan 29 16:12:41 crc kubenswrapper[4714]: I0129 16:12:41.739794 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lz6mw" podUID="7a1dfb55-8680-4cbe-bd78-caca2e847caf" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 16:12:42 crc kubenswrapper[4714]: I0129 16:12:42.740870 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:42 crc kubenswrapper[4714]: I0129 16:12:42.743757 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-lz6mw" Jan 29 16:12:45 crc kubenswrapper[4714]: I0129 16:12:45.717265 4714 patch_prober.go:28] interesting pod/console-f9d7485db-m2g9h container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Jan 29 16:12:45 crc kubenswrapper[4714]: I0129 16:12:45.717607 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-m2g9h" podUID="0e2a789d-6a90-4d60-881e-9562cd92e0a7" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Jan 29 16:12:45 crc kubenswrapper[4714]: I0129 16:12:45.796025 4714 patch_prober.go:28] interesting pod/downloads-7954f5f757-fn75b container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 16:12:45 crc kubenswrapper[4714]: I0129 16:12:45.796079 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-fn75b" podUID="42b66dc3-a385-4350-a943-50f062da35f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 16:12:45 crc kubenswrapper[4714]: I0129 16:12:45.796118 4714 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-fn75b" Jan 29 16:12:45 crc kubenswrapper[4714]: I0129 16:12:45.796288 4714 patch_prober.go:28] interesting pod/downloads-7954f5f757-fn75b container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 16:12:45 crc kubenswrapper[4714]: I0129 16:12:45.796340 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fn75b" podUID="42b66dc3-a385-4350-a943-50f062da35f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 16:12:45 crc kubenswrapper[4714]: I0129 16:12:45.796682 4714 patch_prober.go:28] interesting pod/downloads-7954f5f757-fn75b container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 16:12:45 crc kubenswrapper[4714]: I0129 16:12:45.796724 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fn75b" podUID="42b66dc3-a385-4350-a943-50f062da35f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 16:12:45 crc kubenswrapper[4714]: I0129 16:12:45.796763 4714 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"c4aaec06be7df88764d0dc745049e2e561eb871b6ccb463e86a9ef317a262a34"} pod="openshift-console/downloads-7954f5f757-fn75b" containerMessage="Container download-server failed liveness probe, will be restarted" Jan 29 16:12:45 crc kubenswrapper[4714]: I0129 16:12:45.797739 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-fn75b" podUID="42b66dc3-a385-4350-a943-50f062da35f7" containerName="download-server" containerID="cri-o://c4aaec06be7df88764d0dc745049e2e561eb871b6ccb463e86a9ef317a262a34" gracePeriod=2 Jan 29 16:12:48 crc kubenswrapper[4714]: I0129 16:12:48.680996 4714 generic.go:334] "Generic (PLEG): container finished" podID="42b66dc3-a385-4350-a943-50f062da35f7" containerID="c4aaec06be7df88764d0dc745049e2e561eb871b6ccb463e86a9ef317a262a34" exitCode=0 Jan 29 16:12:48 crc kubenswrapper[4714]: I0129 16:12:48.681089 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-fn75b" event={"ID":"42b66dc3-a385-4350-a943-50f062da35f7","Type":"ContainerDied","Data":"c4aaec06be7df88764d0dc745049e2e561eb871b6ccb463e86a9ef317a262a34"} Jan 29 16:12:52 crc kubenswrapper[4714]: I0129 16:12:52.030735 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:12:55 crc kubenswrapper[4714]: I0129 16:12:55.745431 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:55 crc kubenswrapper[4714]: I0129 16:12:55.751637 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-m2g9h" Jan 29 16:12:55 crc kubenswrapper[4714]: I0129 16:12:55.800509 4714 patch_prober.go:28] interesting pod/downloads-7954f5f757-fn75b container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 16:12:55 crc kubenswrapper[4714]: I0129 16:12:55.800593 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fn75b" podUID="42b66dc3-a385-4350-a943-50f062da35f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 16:12:56 crc kubenswrapper[4714]: I0129 16:12:56.892577 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pbrmk" Jan 29 16:12:57 crc kubenswrapper[4714]: I0129 16:12:57.844721 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:12:57 crc kubenswrapper[4714]: I0129 16:12:57.845249 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:13:03 crc kubenswrapper[4714]: I0129 16:13:03.046092 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 16:13:05 crc kubenswrapper[4714]: I0129 16:13:05.795252 4714 patch_prober.go:28] interesting pod/downloads-7954f5f757-fn75b container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 16:13:05 crc kubenswrapper[4714]: I0129 16:13:05.795325 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fn75b" podUID="42b66dc3-a385-4350-a943-50f062da35f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 16:13:06 crc kubenswrapper[4714]: I0129 16:13:06.818150 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 29 16:13:06 crc kubenswrapper[4714]: E0129 16:13:06.818758 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58" containerName="pruner" Jan 29 16:13:06 crc kubenswrapper[4714]: I0129 16:13:06.818774 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58" containerName="pruner" Jan 29 16:13:06 crc kubenswrapper[4714]: I0129 16:13:06.818923 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="01f3f98d-08b5-47f1-b0a9-8bdb0ad67d58" containerName="pruner" Jan 29 16:13:06 crc kubenswrapper[4714]: I0129 16:13:06.819435 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 16:13:06 crc kubenswrapper[4714]: I0129 16:13:06.823359 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 29 16:13:06 crc kubenswrapper[4714]: I0129 16:13:06.823726 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 29 16:13:06 crc kubenswrapper[4714]: I0129 16:13:06.832059 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 29 16:13:06 crc kubenswrapper[4714]: I0129 16:13:06.921185 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 16:13:06 crc kubenswrapper[4714]: I0129 16:13:06.921236 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 16:13:07 crc kubenswrapper[4714]: I0129 16:13:07.022588 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 16:13:07 crc kubenswrapper[4714]: I0129 16:13:07.022670 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 16:13:07 crc kubenswrapper[4714]: I0129 16:13:07.022764 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 16:13:07 crc kubenswrapper[4714]: I0129 16:13:07.057455 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 16:13:07 crc kubenswrapper[4714]: I0129 16:13:07.198159 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 16:13:08 crc kubenswrapper[4714]: E0129 16:13:08.419084 4714 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \"/var/tmp/container_images_storage1649454453/4\": happened during read: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 16:13:08 crc kubenswrapper[4714]: E0129 16:13:08.419572 4714 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gks4z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-6bjgq_openshift-marketplace(98a35d03-ef3b-4341-9866-56d12a28aee3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \"/var/tmp/container_images_storage1649454453/4\": happened during read: context canceled" logger="UnhandledError" Jan 29 16:13:08 crc kubenswrapper[4714]: E0129 16:13:08.421724 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \\\"/var/tmp/container_images_storage1649454453/4\\\": happened during read: context canceled\"" pod="openshift-marketplace/community-operators-6bjgq" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" Jan 29 16:13:11 crc kubenswrapper[4714]: E0129 16:13:11.753036 4714 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 16:13:11 crc kubenswrapper[4714]: E0129 16:13:11.753227 4714 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9dv2v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-dbvgp_openshift-marketplace(ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 16:13:11 crc kubenswrapper[4714]: E0129 16:13:11.754443 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-dbvgp" podUID="ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef" Jan 29 16:13:12 crc kubenswrapper[4714]: I0129 16:13:12.012866 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 29 16:13:12 crc kubenswrapper[4714]: I0129 16:13:12.013740 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 29 16:13:12 crc kubenswrapper[4714]: I0129 16:13:12.037478 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 29 16:13:12 crc kubenswrapper[4714]: I0129 16:13:12.126714 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24-kube-api-access\") pod \"installer-9-crc\" (UID: \"9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 16:13:12 crc kubenswrapper[4714]: I0129 16:13:12.127063 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24-kubelet-dir\") pod \"installer-9-crc\" (UID: \"9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 16:13:12 crc kubenswrapper[4714]: I0129 16:13:12.127252 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24-var-lock\") pod \"installer-9-crc\" (UID: \"9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 16:13:12 crc kubenswrapper[4714]: I0129 16:13:12.228752 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24-kubelet-dir\") pod \"installer-9-crc\" (UID: \"9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 16:13:12 crc kubenswrapper[4714]: I0129 16:13:12.228801 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24-var-lock\") pod \"installer-9-crc\" (UID: \"9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 16:13:12 crc kubenswrapper[4714]: I0129 16:13:12.228837 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24-kube-api-access\") pod \"installer-9-crc\" (UID: \"9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 16:13:12 crc kubenswrapper[4714]: I0129 16:13:12.228913 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24-var-lock\") pod \"installer-9-crc\" (UID: \"9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 16:13:12 crc kubenswrapper[4714]: I0129 16:13:12.229004 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24-kubelet-dir\") pod \"installer-9-crc\" (UID: \"9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 16:13:12 crc kubenswrapper[4714]: I0129 16:13:12.245984 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24-kube-api-access\") pod \"installer-9-crc\" (UID: \"9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 16:13:12 crc kubenswrapper[4714]: E0129 16:13:12.328112 4714 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 16:13:12 crc kubenswrapper[4714]: E0129 16:13:12.328271 4714 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v8bfv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-74twj_openshift-marketplace(a97ed1ff-657f-4bde-943b-78caf9d07f92): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 16:13:12 crc kubenswrapper[4714]: E0129 16:13:12.329826 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-74twj" podUID="a97ed1ff-657f-4bde-943b-78caf9d07f92" Jan 29 16:13:12 crc kubenswrapper[4714]: I0129 16:13:12.345662 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 29 16:13:12 crc kubenswrapper[4714]: E0129 16:13:12.642136 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-6bjgq" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" Jan 29 16:13:12 crc kubenswrapper[4714]: E0129 16:13:12.642174 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-dbvgp" podUID="ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef" Jan 29 16:13:12 crc kubenswrapper[4714]: E0129 16:13:12.754346 4714 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 16:13:12 crc kubenswrapper[4714]: E0129 16:13:12.754541 4714 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m4jjj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-lb68h_openshift-marketplace(d05e7c79-7d66-4453-aedb-f240784ff294): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 16:13:12 crc kubenswrapper[4714]: E0129 16:13:12.755734 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-lb68h" podUID="d05e7c79-7d66-4453-aedb-f240784ff294" Jan 29 16:13:13 crc kubenswrapper[4714]: E0129 16:13:13.494314 4714 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:94d1bfc77428a945334e81bab025286e1fb0c1323b3aa1395b0c2f8e42153686: Get \"https://registry.redhat.io/v2/redhat/community-operator-index/blobs/sha256:94d1bfc77428a945334e81bab025286e1fb0c1323b3aa1395b0c2f8e42153686\": context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 16:13:13 crc kubenswrapper[4714]: E0129 16:13:13.494622 4714 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zbn22,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-xtr82_openshift-marketplace(11a30de8-b234-47b4-8fd0-44f0c428be78): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:94d1bfc77428a945334e81bab025286e1fb0c1323b3aa1395b0c2f8e42153686: Get \"https://registry.redhat.io/v2/redhat/community-operator-index/blobs/sha256:94d1bfc77428a945334e81bab025286e1fb0c1323b3aa1395b0c2f8e42153686\": context canceled" logger="UnhandledError" Jan 29 16:13:13 crc kubenswrapper[4714]: E0129 16:13:13.496439 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:94d1bfc77428a945334e81bab025286e1fb0c1323b3aa1395b0c2f8e42153686: Get \\\"https://registry.redhat.io/v2/redhat/community-operator-index/blobs/sha256:94d1bfc77428a945334e81bab025286e1fb0c1323b3aa1395b0c2f8e42153686\\\": context canceled\"" pod="openshift-marketplace/community-operators-xtr82" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" Jan 29 16:13:15 crc kubenswrapper[4714]: I0129 16:13:15.797170 4714 patch_prober.go:28] interesting pod/downloads-7954f5f757-fn75b container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 16:13:15 crc kubenswrapper[4714]: I0129 16:13:15.797277 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fn75b" podUID="42b66dc3-a385-4350-a943-50f062da35f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 16:13:16 crc kubenswrapper[4714]: E0129 16:13:16.356700 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-74twj" podUID="a97ed1ff-657f-4bde-943b-78caf9d07f92" Jan 29 16:13:16 crc kubenswrapper[4714]: E0129 16:13:16.357927 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-lb68h" podUID="d05e7c79-7d66-4453-aedb-f240784ff294" Jan 29 16:13:16 crc kubenswrapper[4714]: E0129 16:13:16.357889 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xtr82" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" Jan 29 16:13:16 crc kubenswrapper[4714]: E0129 16:13:16.397169 4714 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 16:13:16 crc kubenswrapper[4714]: E0129 16:13:16.397334 4714 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xpzw5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-kg9qt_openshift-marketplace(ec0eba7e-2ea0-432f-bc57-d87404801abe): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 16:13:16 crc kubenswrapper[4714]: E0129 16:13:16.398718 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-kg9qt" podUID="ec0eba7e-2ea0-432f-bc57-d87404801abe" Jan 29 16:13:16 crc kubenswrapper[4714]: E0129 16:13:16.430723 4714 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 16:13:16 crc kubenswrapper[4714]: E0129 16:13:16.430873 4714 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ntklh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-nssrv_openshift-marketplace(eae853ba-61c9-439b-9dc9-21567075f18a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 16:13:16 crc kubenswrapper[4714]: E0129 16:13:16.432132 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-nssrv" podUID="eae853ba-61c9-439b-9dc9-21567075f18a" Jan 29 16:13:16 crc kubenswrapper[4714]: E0129 16:13:16.444108 4714 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 16:13:16 crc kubenswrapper[4714]: E0129 16:13:16.444299 4714 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8pkm4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-hbcnj_openshift-marketplace(213a402c-b327-4aa6-9690-22d6da8664a4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 16:13:16 crc kubenswrapper[4714]: E0129 16:13:16.445671 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-hbcnj" podUID="213a402c-b327-4aa6-9690-22d6da8664a4" Jan 29 16:13:16 crc kubenswrapper[4714]: I0129 16:13:16.838219 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 29 16:13:16 crc kubenswrapper[4714]: I0129 16:13:16.867528 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-fn75b" event={"ID":"42b66dc3-a385-4350-a943-50f062da35f7","Type":"ContainerStarted","Data":"d3f1419598c4245050212acd4731aa1926279689462b7e83b132ac3c6307471d"} Jan 29 16:13:16 crc kubenswrapper[4714]: I0129 16:13:16.868407 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-fn75b" Jan 29 16:13:16 crc kubenswrapper[4714]: I0129 16:13:16.868472 4714 patch_prober.go:28] interesting pod/downloads-7954f5f757-fn75b container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 16:13:16 crc kubenswrapper[4714]: I0129 16:13:16.868496 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fn75b" podUID="42b66dc3-a385-4350-a943-50f062da35f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 16:13:16 crc kubenswrapper[4714]: I0129 16:13:16.872555 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24","Type":"ContainerStarted","Data":"c9979ec9683e16f0b3aadc58f285988aa9a20624d0e8ae1b706d16b4bb36c291"} Jan 29 16:13:16 crc kubenswrapper[4714]: I0129 16:13:16.877520 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2w92b" event={"ID":"791456e8-8d95-4cdb-8fd1-d06a7586b328","Type":"ContainerStarted","Data":"e44ed67e5705bbd4a9aa4ca26f7b537b9562b4aa728317770efcfa94c9759009"} Jan 29 16:13:16 crc kubenswrapper[4714]: I0129 16:13:16.910657 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 29 16:13:16 crc kubenswrapper[4714]: I0129 16:13:16.929476 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-2w92b" podStartSLOduration=182.929444434 podStartE2EDuration="3m2.929444434s" podCreationTimestamp="2026-01-29 16:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:13:16.929401543 +0000 UTC m=+203.449902663" watchObservedRunningTime="2026-01-29 16:13:16.929444434 +0000 UTC m=+203.449945554" Jan 29 16:13:17 crc kubenswrapper[4714]: I0129 16:13:17.886796 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb","Type":"ContainerStarted","Data":"fea6462f2633259cc5627c9d340316b80f17acfb1fb2132675b377437533a003"} Jan 29 16:13:17 crc kubenswrapper[4714]: I0129 16:13:17.887752 4714 patch_prober.go:28] interesting pod/downloads-7954f5f757-fn75b container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 16:13:17 crc kubenswrapper[4714]: I0129 16:13:17.887823 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fn75b" podUID="42b66dc3-a385-4350-a943-50f062da35f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 16:13:18 crc kubenswrapper[4714]: I0129 16:13:18.898883 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24","Type":"ContainerStarted","Data":"d7f3c5ef1cf90e64e41a56b19a6b215a0e1a7265f9a460a77da807a32640c2c8"} Jan 29 16:13:18 crc kubenswrapper[4714]: I0129 16:13:18.903872 4714 patch_prober.go:28] interesting pod/downloads-7954f5f757-fn75b container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 16:13:18 crc kubenswrapper[4714]: I0129 16:13:18.904015 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fn75b" podUID="42b66dc3-a385-4350-a943-50f062da35f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 16:13:18 crc kubenswrapper[4714]: I0129 16:13:18.904467 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb","Type":"ContainerStarted","Data":"f4c0a8611fe19893c08801182acc09882f18c15576bc5e70f9010e8259dff15b"} Jan 29 16:13:18 crc kubenswrapper[4714]: I0129 16:13:18.933348 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=6.933321122 podStartE2EDuration="6.933321122s" podCreationTimestamp="2026-01-29 16:13:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:13:18.926207718 +0000 UTC m=+205.446708878" watchObservedRunningTime="2026-01-29 16:13:18.933321122 +0000 UTC m=+205.453822282" Jan 29 16:13:18 crc kubenswrapper[4714]: I0129 16:13:18.959189 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=12.9591554 podStartE2EDuration="12.9591554s" podCreationTimestamp="2026-01-29 16:13:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:13:18.953165729 +0000 UTC m=+205.473666859" watchObservedRunningTime="2026-01-29 16:13:18.9591554 +0000 UTC m=+205.479656530" Jan 29 16:13:19 crc kubenswrapper[4714]: I0129 16:13:19.915199 4714 generic.go:334] "Generic (PLEG): container finished" podID="c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb" containerID="f4c0a8611fe19893c08801182acc09882f18c15576bc5e70f9010e8259dff15b" exitCode=0 Jan 29 16:13:19 crc kubenswrapper[4714]: I0129 16:13:19.915450 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb","Type":"ContainerDied","Data":"f4c0a8611fe19893c08801182acc09882f18c15576bc5e70f9010e8259dff15b"} Jan 29 16:13:21 crc kubenswrapper[4714]: I0129 16:13:21.202714 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 16:13:21 crc kubenswrapper[4714]: I0129 16:13:21.381066 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb-kube-api-access\") pod \"c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb\" (UID: \"c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb\") " Jan 29 16:13:21 crc kubenswrapper[4714]: I0129 16:13:21.382119 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb-kubelet-dir\") pod \"c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb\" (UID: \"c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb\") " Jan 29 16:13:21 crc kubenswrapper[4714]: I0129 16:13:21.382348 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb" (UID: "c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:13:21 crc kubenswrapper[4714]: I0129 16:13:21.388958 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb" (UID: "c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:13:21 crc kubenswrapper[4714]: I0129 16:13:21.483850 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 16:13:21 crc kubenswrapper[4714]: I0129 16:13:21.483902 4714 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 29 16:13:21 crc kubenswrapper[4714]: I0129 16:13:21.932450 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb","Type":"ContainerDied","Data":"fea6462f2633259cc5627c9d340316b80f17acfb1fb2132675b377437533a003"} Jan 29 16:13:21 crc kubenswrapper[4714]: I0129 16:13:21.932524 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fea6462f2633259cc5627c9d340316b80f17acfb1fb2132675b377437533a003" Jan 29 16:13:21 crc kubenswrapper[4714]: I0129 16:13:21.932590 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 16:13:25 crc kubenswrapper[4714]: I0129 16:13:25.796228 4714 patch_prober.go:28] interesting pod/downloads-7954f5f757-fn75b container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 16:13:25 crc kubenswrapper[4714]: I0129 16:13:25.796644 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fn75b" podUID="42b66dc3-a385-4350-a943-50f062da35f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 16:13:25 crc kubenswrapper[4714]: I0129 16:13:25.796265 4714 patch_prober.go:28] interesting pod/downloads-7954f5f757-fn75b container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 16:13:25 crc kubenswrapper[4714]: I0129 16:13:25.796766 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-fn75b" podUID="42b66dc3-a385-4350-a943-50f062da35f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 16:13:27 crc kubenswrapper[4714]: I0129 16:13:27.844092 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:13:27 crc kubenswrapper[4714]: I0129 16:13:27.844529 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:13:27 crc kubenswrapper[4714]: I0129 16:13:27.844687 4714 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:13:27 crc kubenswrapper[4714]: I0129 16:13:27.845749 4714 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5"} pod="openshift-machine-config-operator/machine-config-daemon-ppngk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 16:13:27 crc kubenswrapper[4714]: I0129 16:13:27.845882 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" containerID="cri-o://27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5" gracePeriod=600 Jan 29 16:13:29 crc kubenswrapper[4714]: I0129 16:13:29.987095 4714 generic.go:334] "Generic (PLEG): container finished" podID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerID="27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5" exitCode=0 Jan 29 16:13:29 crc kubenswrapper[4714]: I0129 16:13:29.987156 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" event={"ID":"c8c765f3-89eb-4077-8829-03e86eb0c90c","Type":"ContainerDied","Data":"27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5"} Jan 29 16:13:35 crc kubenswrapper[4714]: I0129 16:13:35.029950 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" event={"ID":"c8c765f3-89eb-4077-8829-03e86eb0c90c","Type":"ContainerStarted","Data":"32d59d2a4eb095db60ac3365411265035b47b1f01d164e950c86daa5aecb2792"} Jan 29 16:13:35 crc kubenswrapper[4714]: I0129 16:13:35.038827 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbvgp" event={"ID":"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef","Type":"ContainerDied","Data":"e4723c2a15106e7cb0b50986860a7a558bd1c3ece5b50a3175d026187fce0bd7"} Jan 29 16:13:35 crc kubenswrapper[4714]: I0129 16:13:35.031766 4714 generic.go:334] "Generic (PLEG): container finished" podID="ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef" containerID="e4723c2a15106e7cb0b50986860a7a558bd1c3ece5b50a3175d026187fce0bd7" exitCode=0 Jan 29 16:13:35 crc kubenswrapper[4714]: I0129 16:13:35.819248 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-fn75b" Jan 29 16:13:36 crc kubenswrapper[4714]: I0129 16:13:36.050655 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg9qt" event={"ID":"ec0eba7e-2ea0-432f-bc57-d87404801abe","Type":"ContainerStarted","Data":"70619171b98977a9d286bd513ce901776ee376ec0c300ee22dba9b847da868b1"} Jan 29 16:13:36 crc kubenswrapper[4714]: I0129 16:13:36.052895 4714 generic.go:334] "Generic (PLEG): container finished" podID="213a402c-b327-4aa6-9690-22d6da8664a4" containerID="30c0abcdea89377220aa9b27f7c4d0c14e091ab28605ec484a71ffefcd4d36ba" exitCode=0 Jan 29 16:13:36 crc kubenswrapper[4714]: I0129 16:13:36.052987 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hbcnj" event={"ID":"213a402c-b327-4aa6-9690-22d6da8664a4","Type":"ContainerDied","Data":"30c0abcdea89377220aa9b27f7c4d0c14e091ab28605ec484a71ffefcd4d36ba"} Jan 29 16:13:36 crc kubenswrapper[4714]: I0129 16:13:36.058728 4714 generic.go:334] "Generic (PLEG): container finished" podID="a97ed1ff-657f-4bde-943b-78caf9d07f92" containerID="b55a4a53101476467f28cefb11eb7b554ba1847628876fc42f62a75c8730a4f6" exitCode=0 Jan 29 16:13:36 crc kubenswrapper[4714]: I0129 16:13:36.058823 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-74twj" event={"ID":"a97ed1ff-657f-4bde-943b-78caf9d07f92","Type":"ContainerDied","Data":"b55a4a53101476467f28cefb11eb7b554ba1847628876fc42f62a75c8730a4f6"} Jan 29 16:13:36 crc kubenswrapper[4714]: I0129 16:13:36.062480 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbvgp" event={"ID":"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef","Type":"ContainerStarted","Data":"4d3b6ad01b47ee4415724a7f3a339855145f38f464bbba6afefebc62453be6ce"} Jan 29 16:13:36 crc kubenswrapper[4714]: I0129 16:13:36.064918 4714 generic.go:334] "Generic (PLEG): container finished" podID="d05e7c79-7d66-4453-aedb-f240784ff294" containerID="dc4ea141b6a80ae098de7ca5a17e9e2b1ec3ebf2112f640c6fad4d5fcc75a51c" exitCode=0 Jan 29 16:13:36 crc kubenswrapper[4714]: I0129 16:13:36.065028 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lb68h" event={"ID":"d05e7c79-7d66-4453-aedb-f240784ff294","Type":"ContainerDied","Data":"dc4ea141b6a80ae098de7ca5a17e9e2b1ec3ebf2112f640c6fad4d5fcc75a51c"} Jan 29 16:13:36 crc kubenswrapper[4714]: I0129 16:13:36.075241 4714 generic.go:334] "Generic (PLEG): container finished" podID="eae853ba-61c9-439b-9dc9-21567075f18a" containerID="af26cf3e86e4001d9e7f83e8aa3b6ea940d03f589aa1d1907c488eb1df46568b" exitCode=0 Jan 29 16:13:36 crc kubenswrapper[4714]: I0129 16:13:36.076257 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nssrv" event={"ID":"eae853ba-61c9-439b-9dc9-21567075f18a","Type":"ContainerDied","Data":"af26cf3e86e4001d9e7f83e8aa3b6ea940d03f589aa1d1907c488eb1df46568b"} Jan 29 16:13:36 crc kubenswrapper[4714]: I0129 16:13:36.108617 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dbvgp" podStartSLOduration=4.12663362 podStartE2EDuration="1m9.108598348s" podCreationTimestamp="2026-01-29 16:12:27 +0000 UTC" firstStartedPulling="2026-01-29 16:12:30.511213659 +0000 UTC m=+157.031714789" lastFinishedPulling="2026-01-29 16:13:35.493178397 +0000 UTC m=+222.013679517" observedRunningTime="2026-01-29 16:13:36.105593762 +0000 UTC m=+222.626094882" watchObservedRunningTime="2026-01-29 16:13:36.108598348 +0000 UTC m=+222.629099458" Jan 29 16:13:37 crc kubenswrapper[4714]: I0129 16:13:37.089283 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lb68h" event={"ID":"d05e7c79-7d66-4453-aedb-f240784ff294","Type":"ContainerStarted","Data":"25adf524f9c5473bfa242fd63827380ee76b9012a8eb32666464c503ff1ae5f5"} Jan 29 16:13:37 crc kubenswrapper[4714]: I0129 16:13:37.093231 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nssrv" event={"ID":"eae853ba-61c9-439b-9dc9-21567075f18a","Type":"ContainerStarted","Data":"52559e4420e032272c3033326a96872ff005d794cf95b2dce2bffa130b6cf392"} Jan 29 16:13:37 crc kubenswrapper[4714]: I0129 16:13:37.095509 4714 generic.go:334] "Generic (PLEG): container finished" podID="ec0eba7e-2ea0-432f-bc57-d87404801abe" containerID="70619171b98977a9d286bd513ce901776ee376ec0c300ee22dba9b847da868b1" exitCode=0 Jan 29 16:13:37 crc kubenswrapper[4714]: I0129 16:13:37.095543 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg9qt" event={"ID":"ec0eba7e-2ea0-432f-bc57-d87404801abe","Type":"ContainerDied","Data":"70619171b98977a9d286bd513ce901776ee376ec0c300ee22dba9b847da868b1"} Jan 29 16:13:37 crc kubenswrapper[4714]: I0129 16:13:37.095557 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg9qt" event={"ID":"ec0eba7e-2ea0-432f-bc57-d87404801abe","Type":"ContainerStarted","Data":"280fd3af9e55869b591af38324a50ac2f821ce1422ef0def509f857f44fa4f86"} Jan 29 16:13:37 crc kubenswrapper[4714]: I0129 16:13:37.101420 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hbcnj" event={"ID":"213a402c-b327-4aa6-9690-22d6da8664a4","Type":"ContainerStarted","Data":"2c317c97caab5fe2de4c3ba52a485756258217cc8c0f8aff08364a47758a8598"} Jan 29 16:13:37 crc kubenswrapper[4714]: I0129 16:13:37.103427 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-74twj" event={"ID":"a97ed1ff-657f-4bde-943b-78caf9d07f92","Type":"ContainerStarted","Data":"0725e5fe8581c9f5ab88fa8ad4af11d5f996f4972602031ec88175b5fda32f79"} Jan 29 16:13:37 crc kubenswrapper[4714]: I0129 16:13:37.111773 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lb68h" podStartSLOduration=4.00847504 podStartE2EDuration="1m8.111761981s" podCreationTimestamp="2026-01-29 16:12:29 +0000 UTC" firstStartedPulling="2026-01-29 16:12:32.553203997 +0000 UTC m=+159.073705117" lastFinishedPulling="2026-01-29 16:13:36.656490938 +0000 UTC m=+223.176992058" observedRunningTime="2026-01-29 16:13:37.110128805 +0000 UTC m=+223.630629925" watchObservedRunningTime="2026-01-29 16:13:37.111761981 +0000 UTC m=+223.632263101" Jan 29 16:13:37 crc kubenswrapper[4714]: I0129 16:13:37.128305 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kg9qt" podStartSLOduration=3.911234839 podStartE2EDuration="1m7.128284414s" podCreationTimestamp="2026-01-29 16:12:30 +0000 UTC" firstStartedPulling="2026-01-29 16:12:33.568827221 +0000 UTC m=+160.089328351" lastFinishedPulling="2026-01-29 16:13:36.785876796 +0000 UTC m=+223.306377926" observedRunningTime="2026-01-29 16:13:37.126827742 +0000 UTC m=+223.647328862" watchObservedRunningTime="2026-01-29 16:13:37.128284414 +0000 UTC m=+223.648785524" Jan 29 16:13:37 crc kubenswrapper[4714]: I0129 16:13:37.145835 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nssrv" podStartSLOduration=3.9601176909999998 podStartE2EDuration="1m9.145819895s" podCreationTimestamp="2026-01-29 16:12:28 +0000 UTC" firstStartedPulling="2026-01-29 16:12:31.553403336 +0000 UTC m=+158.073904456" lastFinishedPulling="2026-01-29 16:13:36.73910554 +0000 UTC m=+223.259606660" observedRunningTime="2026-01-29 16:13:37.14423724 +0000 UTC m=+223.664738360" watchObservedRunningTime="2026-01-29 16:13:37.145819895 +0000 UTC m=+223.666321005" Jan 29 16:13:37 crc kubenswrapper[4714]: I0129 16:13:37.162552 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hbcnj" podStartSLOduration=3.174430813 podStartE2EDuration="1m8.162538483s" podCreationTimestamp="2026-01-29 16:12:29 +0000 UTC" firstStartedPulling="2026-01-29 16:12:31.546307129 +0000 UTC m=+158.066808249" lastFinishedPulling="2026-01-29 16:13:36.534414799 +0000 UTC m=+223.054915919" observedRunningTime="2026-01-29 16:13:37.161887854 +0000 UTC m=+223.682388974" watchObservedRunningTime="2026-01-29 16:13:37.162538483 +0000 UTC m=+223.683039603" Jan 29 16:13:37 crc kubenswrapper[4714]: I0129 16:13:37.184281 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-74twj" podStartSLOduration=5.187567452 podStartE2EDuration="1m11.184265144s" podCreationTimestamp="2026-01-29 16:12:26 +0000 UTC" firstStartedPulling="2026-01-29 16:12:30.511137197 +0000 UTC m=+157.031638337" lastFinishedPulling="2026-01-29 16:13:36.507834909 +0000 UTC m=+223.028336029" observedRunningTime="2026-01-29 16:13:37.180411814 +0000 UTC m=+223.700912954" watchObservedRunningTime="2026-01-29 16:13:37.184265144 +0000 UTC m=+223.704766254" Jan 29 16:13:37 crc kubenswrapper[4714]: I0129 16:13:37.340470 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-74twj" Jan 29 16:13:37 crc kubenswrapper[4714]: I0129 16:13:37.340532 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-74twj" Jan 29 16:13:37 crc kubenswrapper[4714]: I0129 16:13:37.701420 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dbvgp" Jan 29 16:13:37 crc kubenswrapper[4714]: I0129 16:13:37.701755 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dbvgp" Jan 29 16:13:38 crc kubenswrapper[4714]: I0129 16:13:38.522635 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-74twj" podUID="a97ed1ff-657f-4bde-943b-78caf9d07f92" containerName="registry-server" probeResult="failure" output=< Jan 29 16:13:38 crc kubenswrapper[4714]: timeout: failed to connect service ":50051" within 1s Jan 29 16:13:38 crc kubenswrapper[4714]: > Jan 29 16:13:38 crc kubenswrapper[4714]: I0129 16:13:38.737979 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-dbvgp" podUID="ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef" containerName="registry-server" probeResult="failure" output=< Jan 29 16:13:38 crc kubenswrapper[4714]: timeout: failed to connect service ":50051" within 1s Jan 29 16:13:38 crc kubenswrapper[4714]: > Jan 29 16:13:39 crc kubenswrapper[4714]: I0129 16:13:39.063470 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nssrv" Jan 29 16:13:39 crc kubenswrapper[4714]: I0129 16:13:39.063809 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nssrv" Jan 29 16:13:39 crc kubenswrapper[4714]: I0129 16:13:39.116261 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nssrv" Jan 29 16:13:39 crc kubenswrapper[4714]: I0129 16:13:39.463648 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hbcnj" Jan 29 16:13:39 crc kubenswrapper[4714]: I0129 16:13:39.463700 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hbcnj" Jan 29 16:13:39 crc kubenswrapper[4714]: I0129 16:13:39.513564 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hbcnj" Jan 29 16:13:40 crc kubenswrapper[4714]: I0129 16:13:40.263314 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lb68h" Jan 29 16:13:40 crc kubenswrapper[4714]: I0129 16:13:40.264710 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lb68h" Jan 29 16:13:40 crc kubenswrapper[4714]: I0129 16:13:40.706304 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kg9qt" Jan 29 16:13:40 crc kubenswrapper[4714]: I0129 16:13:40.706863 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kg9qt" Jan 29 16:13:41 crc kubenswrapper[4714]: I0129 16:13:41.307021 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lb68h" podUID="d05e7c79-7d66-4453-aedb-f240784ff294" containerName="registry-server" probeResult="failure" output=< Jan 29 16:13:41 crc kubenswrapper[4714]: timeout: failed to connect service ":50051" within 1s Jan 29 16:13:41 crc kubenswrapper[4714]: > Jan 29 16:13:41 crc kubenswrapper[4714]: I0129 16:13:41.758184 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kg9qt" podUID="ec0eba7e-2ea0-432f-bc57-d87404801abe" containerName="registry-server" probeResult="failure" output=< Jan 29 16:13:41 crc kubenswrapper[4714]: timeout: failed to connect service ":50051" within 1s Jan 29 16:13:41 crc kubenswrapper[4714]: > Jan 29 16:13:44 crc kubenswrapper[4714]: I0129 16:13:44.151380 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xtr82" event={"ID":"11a30de8-b234-47b4-8fd0-44f0c428be78","Type":"ContainerStarted","Data":"8a54a45940dea4f60292ad1cc53a3fa404d30031632cdd928c54b5c498a0d947"} Jan 29 16:13:44 crc kubenswrapper[4714]: I0129 16:13:44.156264 4714 generic.go:334] "Generic (PLEG): container finished" podID="98a35d03-ef3b-4341-9866-56d12a28aee3" containerID="aaa814f95896ac1eaedc7b86e2197c8140345f5296d9cb5ecb6a3dac9cee5ab6" exitCode=0 Jan 29 16:13:44 crc kubenswrapper[4714]: I0129 16:13:44.156312 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bjgq" event={"ID":"98a35d03-ef3b-4341-9866-56d12a28aee3","Type":"ContainerDied","Data":"aaa814f95896ac1eaedc7b86e2197c8140345f5296d9cb5ecb6a3dac9cee5ab6"} Jan 29 16:13:45 crc kubenswrapper[4714]: I0129 16:13:45.168689 4714 generic.go:334] "Generic (PLEG): container finished" podID="11a30de8-b234-47b4-8fd0-44f0c428be78" containerID="8a54a45940dea4f60292ad1cc53a3fa404d30031632cdd928c54b5c498a0d947" exitCode=0 Jan 29 16:13:45 crc kubenswrapper[4714]: I0129 16:13:45.168845 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xtr82" event={"ID":"11a30de8-b234-47b4-8fd0-44f0c428be78","Type":"ContainerDied","Data":"8a54a45940dea4f60292ad1cc53a3fa404d30031632cdd928c54b5c498a0d947"} Jan 29 16:13:45 crc kubenswrapper[4714]: I0129 16:13:45.175748 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bjgq" event={"ID":"98a35d03-ef3b-4341-9866-56d12a28aee3","Type":"ContainerStarted","Data":"7d09f4f35a658993558ef15a141e74b57af657ad6538fc1eecadfe107e507595"} Jan 29 16:13:45 crc kubenswrapper[4714]: I0129 16:13:45.217267 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6bjgq" podStartSLOduration=4.146413642 podStartE2EDuration="1m18.21723727s" podCreationTimestamp="2026-01-29 16:12:27 +0000 UTC" firstStartedPulling="2026-01-29 16:12:30.509961651 +0000 UTC m=+157.030462781" lastFinishedPulling="2026-01-29 16:13:44.580785289 +0000 UTC m=+231.101286409" observedRunningTime="2026-01-29 16:13:45.214593364 +0000 UTC m=+231.735094524" watchObservedRunningTime="2026-01-29 16:13:45.21723727 +0000 UTC m=+231.737738430" Jan 29 16:13:46 crc kubenswrapper[4714]: I0129 16:13:46.194143 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xtr82" event={"ID":"11a30de8-b234-47b4-8fd0-44f0c428be78","Type":"ContainerStarted","Data":"a2c143c1c72b06e6085dcb21f057a3e45d817ac3a8bf8d7e3516db54610f130e"} Jan 29 16:13:46 crc kubenswrapper[4714]: I0129 16:13:46.214691 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xtr82" podStartSLOduration=4.97714805 podStartE2EDuration="1m20.21466974s" podCreationTimestamp="2026-01-29 16:12:26 +0000 UTC" firstStartedPulling="2026-01-29 16:12:30.527830307 +0000 UTC m=+157.048331427" lastFinishedPulling="2026-01-29 16:13:45.765351997 +0000 UTC m=+232.285853117" observedRunningTime="2026-01-29 16:13:46.213429924 +0000 UTC m=+232.733931034" watchObservedRunningTime="2026-01-29 16:13:46.21466974 +0000 UTC m=+232.735170860" Jan 29 16:13:47 crc kubenswrapper[4714]: I0129 16:13:47.062528 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xtr82" Jan 29 16:13:47 crc kubenswrapper[4714]: I0129 16:13:47.062784 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xtr82" Jan 29 16:13:47 crc kubenswrapper[4714]: I0129 16:13:47.401755 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-74twj" Jan 29 16:13:47 crc kubenswrapper[4714]: I0129 16:13:47.447305 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-74twj" Jan 29 16:13:47 crc kubenswrapper[4714]: I0129 16:13:47.528302 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6bjgq" Jan 29 16:13:47 crc kubenswrapper[4714]: I0129 16:13:47.528355 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6bjgq" Jan 29 16:13:47 crc kubenswrapper[4714]: I0129 16:13:47.563418 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6bjgq" Jan 29 16:13:47 crc kubenswrapper[4714]: I0129 16:13:47.749881 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dbvgp" Jan 29 16:13:47 crc kubenswrapper[4714]: I0129 16:13:47.805554 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dbvgp" Jan 29 16:13:48 crc kubenswrapper[4714]: I0129 16:13:48.125316 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-xtr82" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" containerName="registry-server" probeResult="failure" output=< Jan 29 16:13:48 crc kubenswrapper[4714]: timeout: failed to connect service ":50051" within 1s Jan 29 16:13:48 crc kubenswrapper[4714]: > Jan 29 16:13:49 crc kubenswrapper[4714]: I0129 16:13:49.119762 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nssrv" Jan 29 16:13:49 crc kubenswrapper[4714]: I0129 16:13:49.192597 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dbvgp"] Jan 29 16:13:49 crc kubenswrapper[4714]: I0129 16:13:49.208180 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dbvgp" podUID="ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef" containerName="registry-server" containerID="cri-o://4d3b6ad01b47ee4415724a7f3a339855145f38f464bbba6afefebc62453be6ce" gracePeriod=2 Jan 29 16:13:49 crc kubenswrapper[4714]: I0129 16:13:49.544287 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hbcnj" Jan 29 16:13:50 crc kubenswrapper[4714]: I0129 16:13:50.216574 4714 generic.go:334] "Generic (PLEG): container finished" podID="ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef" containerID="4d3b6ad01b47ee4415724a7f3a339855145f38f464bbba6afefebc62453be6ce" exitCode=0 Jan 29 16:13:50 crc kubenswrapper[4714]: I0129 16:13:50.216611 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbvgp" event={"ID":"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef","Type":"ContainerDied","Data":"4d3b6ad01b47ee4415724a7f3a339855145f38f464bbba6afefebc62453be6ce"} Jan 29 16:13:50 crc kubenswrapper[4714]: I0129 16:13:50.325950 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lb68h" Jan 29 16:13:50 crc kubenswrapper[4714]: I0129 16:13:50.389958 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lb68h" Jan 29 16:13:50 crc kubenswrapper[4714]: I0129 16:13:50.670123 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbvgp" Jan 29 16:13:50 crc kubenswrapper[4714]: I0129 16:13:50.703382 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dv2v\" (UniqueName: \"kubernetes.io/projected/ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef-kube-api-access-9dv2v\") pod \"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef\" (UID: \"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef\") " Jan 29 16:13:50 crc kubenswrapper[4714]: I0129 16:13:50.703454 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef-utilities\") pod \"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef\" (UID: \"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef\") " Jan 29 16:13:50 crc kubenswrapper[4714]: I0129 16:13:50.703851 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef-catalog-content\") pod \"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef\" (UID: \"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef\") " Jan 29 16:13:50 crc kubenswrapper[4714]: I0129 16:13:50.704573 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef-utilities" (OuterVolumeSpecName: "utilities") pod "ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef" (UID: "ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:13:50 crc kubenswrapper[4714]: I0129 16:13:50.711447 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef-kube-api-access-9dv2v" (OuterVolumeSpecName: "kube-api-access-9dv2v") pod "ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef" (UID: "ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef"). InnerVolumeSpecName "kube-api-access-9dv2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:13:50 crc kubenswrapper[4714]: I0129 16:13:50.750182 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef" (UID: "ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:13:50 crc kubenswrapper[4714]: I0129 16:13:50.758326 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kg9qt" Jan 29 16:13:50 crc kubenswrapper[4714]: I0129 16:13:50.801080 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kg9qt" Jan 29 16:13:50 crc kubenswrapper[4714]: I0129 16:13:50.805888 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dv2v\" (UniqueName: \"kubernetes.io/projected/ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef-kube-api-access-9dv2v\") on node \"crc\" DevicePath \"\"" Jan 29 16:13:50 crc kubenswrapper[4714]: I0129 16:13:50.805983 4714 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:13:50 crc kubenswrapper[4714]: I0129 16:13:50.805999 4714 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:13:51 crc kubenswrapper[4714]: I0129 16:13:51.226553 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbvgp" event={"ID":"ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef","Type":"ContainerDied","Data":"8245ba1ef35303ce5087b4ec9f0268e726a450c3c8f0f72042d1655209fffe8b"} Jan 29 16:13:51 crc kubenswrapper[4714]: I0129 16:13:51.226695 4714 scope.go:117] "RemoveContainer" containerID="4d3b6ad01b47ee4415724a7f3a339855145f38f464bbba6afefebc62453be6ce" Jan 29 16:13:51 crc kubenswrapper[4714]: I0129 16:13:51.227171 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbvgp" Jan 29 16:13:51 crc kubenswrapper[4714]: I0129 16:13:51.249892 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dbvgp"] Jan 29 16:13:51 crc kubenswrapper[4714]: I0129 16:13:51.252325 4714 scope.go:117] "RemoveContainer" containerID="e4723c2a15106e7cb0b50986860a7a558bd1c3ece5b50a3175d026187fce0bd7" Jan 29 16:13:51 crc kubenswrapper[4714]: I0129 16:13:51.254216 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dbvgp"] Jan 29 16:13:51 crc kubenswrapper[4714]: I0129 16:13:51.281735 4714 scope.go:117] "RemoveContainer" containerID="bf193fd80bff7bed1bec1edfb59432d0f18ec27217fda44032cd6a47058aee41" Jan 29 16:13:51 crc kubenswrapper[4714]: I0129 16:13:51.585015 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hbcnj"] Jan 29 16:13:51 crc kubenswrapper[4714]: I0129 16:13:51.585229 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hbcnj" podUID="213a402c-b327-4aa6-9690-22d6da8664a4" containerName="registry-server" containerID="cri-o://2c317c97caab5fe2de4c3ba52a485756258217cc8c0f8aff08364a47758a8598" gracePeriod=2 Jan 29 16:13:52 crc kubenswrapper[4714]: I0129 16:13:52.192424 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef" path="/var/lib/kubelet/pods/ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef/volumes" Jan 29 16:13:53 crc kubenswrapper[4714]: I0129 16:13:53.239272 4714 generic.go:334] "Generic (PLEG): container finished" podID="213a402c-b327-4aa6-9690-22d6da8664a4" containerID="2c317c97caab5fe2de4c3ba52a485756258217cc8c0f8aff08364a47758a8598" exitCode=0 Jan 29 16:13:53 crc kubenswrapper[4714]: I0129 16:13:53.239321 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hbcnj" event={"ID":"213a402c-b327-4aa6-9690-22d6da8664a4","Type":"ContainerDied","Data":"2c317c97caab5fe2de4c3ba52a485756258217cc8c0f8aff08364a47758a8598"} Jan 29 16:13:53 crc kubenswrapper[4714]: I0129 16:13:53.538622 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hbcnj" Jan 29 16:13:53 crc kubenswrapper[4714]: I0129 16:13:53.638806 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/213a402c-b327-4aa6-9690-22d6da8664a4-catalog-content\") pod \"213a402c-b327-4aa6-9690-22d6da8664a4\" (UID: \"213a402c-b327-4aa6-9690-22d6da8664a4\") " Jan 29 16:13:53 crc kubenswrapper[4714]: I0129 16:13:53.638905 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/213a402c-b327-4aa6-9690-22d6da8664a4-utilities\") pod \"213a402c-b327-4aa6-9690-22d6da8664a4\" (UID: \"213a402c-b327-4aa6-9690-22d6da8664a4\") " Jan 29 16:13:53 crc kubenswrapper[4714]: I0129 16:13:53.638972 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pkm4\" (UniqueName: \"kubernetes.io/projected/213a402c-b327-4aa6-9690-22d6da8664a4-kube-api-access-8pkm4\") pod \"213a402c-b327-4aa6-9690-22d6da8664a4\" (UID: \"213a402c-b327-4aa6-9690-22d6da8664a4\") " Jan 29 16:13:53 crc kubenswrapper[4714]: I0129 16:13:53.639881 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/213a402c-b327-4aa6-9690-22d6da8664a4-utilities" (OuterVolumeSpecName: "utilities") pod "213a402c-b327-4aa6-9690-22d6da8664a4" (UID: "213a402c-b327-4aa6-9690-22d6da8664a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:13:53 crc kubenswrapper[4714]: I0129 16:13:53.643965 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/213a402c-b327-4aa6-9690-22d6da8664a4-kube-api-access-8pkm4" (OuterVolumeSpecName: "kube-api-access-8pkm4") pod "213a402c-b327-4aa6-9690-22d6da8664a4" (UID: "213a402c-b327-4aa6-9690-22d6da8664a4"). InnerVolumeSpecName "kube-api-access-8pkm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:13:53 crc kubenswrapper[4714]: I0129 16:13:53.660225 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/213a402c-b327-4aa6-9690-22d6da8664a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "213a402c-b327-4aa6-9690-22d6da8664a4" (UID: "213a402c-b327-4aa6-9690-22d6da8664a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:13:53 crc kubenswrapper[4714]: I0129 16:13:53.741046 4714 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/213a402c-b327-4aa6-9690-22d6da8664a4-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:13:53 crc kubenswrapper[4714]: I0129 16:13:53.741070 4714 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/213a402c-b327-4aa6-9690-22d6da8664a4-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:13:53 crc kubenswrapper[4714]: I0129 16:13:53.741080 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pkm4\" (UniqueName: \"kubernetes.io/projected/213a402c-b327-4aa6-9690-22d6da8664a4-kube-api-access-8pkm4\") on node \"crc\" DevicePath \"\"" Jan 29 16:13:53 crc kubenswrapper[4714]: I0129 16:13:53.783296 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kg9qt"] Jan 29 16:13:53 crc kubenswrapper[4714]: I0129 16:13:53.783524 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kg9qt" podUID="ec0eba7e-2ea0-432f-bc57-d87404801abe" containerName="registry-server" containerID="cri-o://280fd3af9e55869b591af38324a50ac2f821ce1422ef0def509f857f44fa4f86" gracePeriod=2 Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.116091 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kg9qt" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.150621 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec0eba7e-2ea0-432f-bc57-d87404801abe-catalog-content\") pod \"ec0eba7e-2ea0-432f-bc57-d87404801abe\" (UID: \"ec0eba7e-2ea0-432f-bc57-d87404801abe\") " Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.150774 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec0eba7e-2ea0-432f-bc57-d87404801abe-utilities\") pod \"ec0eba7e-2ea0-432f-bc57-d87404801abe\" (UID: \"ec0eba7e-2ea0-432f-bc57-d87404801abe\") " Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.151034 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpzw5\" (UniqueName: \"kubernetes.io/projected/ec0eba7e-2ea0-432f-bc57-d87404801abe-kube-api-access-xpzw5\") pod \"ec0eba7e-2ea0-432f-bc57-d87404801abe\" (UID: \"ec0eba7e-2ea0-432f-bc57-d87404801abe\") " Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.152737 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec0eba7e-2ea0-432f-bc57-d87404801abe-utilities" (OuterVolumeSpecName: "utilities") pod "ec0eba7e-2ea0-432f-bc57-d87404801abe" (UID: "ec0eba7e-2ea0-432f-bc57-d87404801abe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.154415 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec0eba7e-2ea0-432f-bc57-d87404801abe-kube-api-access-xpzw5" (OuterVolumeSpecName: "kube-api-access-xpzw5") pod "ec0eba7e-2ea0-432f-bc57-d87404801abe" (UID: "ec0eba7e-2ea0-432f-bc57-d87404801abe"). InnerVolumeSpecName "kube-api-access-xpzw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.246957 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hbcnj" event={"ID":"213a402c-b327-4aa6-9690-22d6da8664a4","Type":"ContainerDied","Data":"cc206117744230d0505236f4ba4b88035d78daaa5da318d54c519b3dd8b10d4e"} Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.247019 4714 scope.go:117] "RemoveContainer" containerID="2c317c97caab5fe2de4c3ba52a485756258217cc8c0f8aff08364a47758a8598" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.247028 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hbcnj" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.251027 4714 generic.go:334] "Generic (PLEG): container finished" podID="ec0eba7e-2ea0-432f-bc57-d87404801abe" containerID="280fd3af9e55869b591af38324a50ac2f821ce1422ef0def509f857f44fa4f86" exitCode=0 Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.251063 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg9qt" event={"ID":"ec0eba7e-2ea0-432f-bc57-d87404801abe","Type":"ContainerDied","Data":"280fd3af9e55869b591af38324a50ac2f821ce1422ef0def509f857f44fa4f86"} Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.251090 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg9qt" event={"ID":"ec0eba7e-2ea0-432f-bc57-d87404801abe","Type":"ContainerDied","Data":"0b72081696a87dc8bb1d6a9fb5d18f85ed353808c0ac8086b9bf62a3458b3736"} Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.251093 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kg9qt" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.252907 4714 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec0eba7e-2ea0-432f-bc57-d87404801abe-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.252924 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpzw5\" (UniqueName: \"kubernetes.io/projected/ec0eba7e-2ea0-432f-bc57-d87404801abe-kube-api-access-xpzw5\") on node \"crc\" DevicePath \"\"" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.269501 4714 scope.go:117] "RemoveContainer" containerID="30c0abcdea89377220aa9b27f7c4d0c14e091ab28605ec484a71ffefcd4d36ba" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.273649 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hbcnj"] Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.274711 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hbcnj"] Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.283281 4714 scope.go:117] "RemoveContainer" containerID="53e99ef17ef7a89695643694f24aab5f9e6445925a8b86fd6d021c08cb6c082f" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.285541 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec0eba7e-2ea0-432f-bc57-d87404801abe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ec0eba7e-2ea0-432f-bc57-d87404801abe" (UID: "ec0eba7e-2ea0-432f-bc57-d87404801abe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.302093 4714 scope.go:117] "RemoveContainer" containerID="280fd3af9e55869b591af38324a50ac2f821ce1422ef0def509f857f44fa4f86" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.316975 4714 scope.go:117] "RemoveContainer" containerID="70619171b98977a9d286bd513ce901776ee376ec0c300ee22dba9b847da868b1" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.332437 4714 scope.go:117] "RemoveContainer" containerID="52ca09e52af15343909984359236a933ffe1f46ab7ada48929cc2741cd10782f" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.351509 4714 scope.go:117] "RemoveContainer" containerID="280fd3af9e55869b591af38324a50ac2f821ce1422ef0def509f857f44fa4f86" Jan 29 16:13:54 crc kubenswrapper[4714]: E0129 16:13:54.351747 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"280fd3af9e55869b591af38324a50ac2f821ce1422ef0def509f857f44fa4f86\": container with ID starting with 280fd3af9e55869b591af38324a50ac2f821ce1422ef0def509f857f44fa4f86 not found: ID does not exist" containerID="280fd3af9e55869b591af38324a50ac2f821ce1422ef0def509f857f44fa4f86" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.351778 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"280fd3af9e55869b591af38324a50ac2f821ce1422ef0def509f857f44fa4f86"} err="failed to get container status \"280fd3af9e55869b591af38324a50ac2f821ce1422ef0def509f857f44fa4f86\": rpc error: code = NotFound desc = could not find container \"280fd3af9e55869b591af38324a50ac2f821ce1422ef0def509f857f44fa4f86\": container with ID starting with 280fd3af9e55869b591af38324a50ac2f821ce1422ef0def509f857f44fa4f86 not found: ID does not exist" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.351803 4714 scope.go:117] "RemoveContainer" containerID="70619171b98977a9d286bd513ce901776ee376ec0c300ee22dba9b847da868b1" Jan 29 16:13:54 crc kubenswrapper[4714]: E0129 16:13:54.352323 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70619171b98977a9d286bd513ce901776ee376ec0c300ee22dba9b847da868b1\": container with ID starting with 70619171b98977a9d286bd513ce901776ee376ec0c300ee22dba9b847da868b1 not found: ID does not exist" containerID="70619171b98977a9d286bd513ce901776ee376ec0c300ee22dba9b847da868b1" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.352345 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70619171b98977a9d286bd513ce901776ee376ec0c300ee22dba9b847da868b1"} err="failed to get container status \"70619171b98977a9d286bd513ce901776ee376ec0c300ee22dba9b847da868b1\": rpc error: code = NotFound desc = could not find container \"70619171b98977a9d286bd513ce901776ee376ec0c300ee22dba9b847da868b1\": container with ID starting with 70619171b98977a9d286bd513ce901776ee376ec0c300ee22dba9b847da868b1 not found: ID does not exist" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.352361 4714 scope.go:117] "RemoveContainer" containerID="52ca09e52af15343909984359236a933ffe1f46ab7ada48929cc2741cd10782f" Jan 29 16:13:54 crc kubenswrapper[4714]: E0129 16:13:54.352563 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52ca09e52af15343909984359236a933ffe1f46ab7ada48929cc2741cd10782f\": container with ID starting with 52ca09e52af15343909984359236a933ffe1f46ab7ada48929cc2741cd10782f not found: ID does not exist" containerID="52ca09e52af15343909984359236a933ffe1f46ab7ada48929cc2741cd10782f" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.352586 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52ca09e52af15343909984359236a933ffe1f46ab7ada48929cc2741cd10782f"} err="failed to get container status \"52ca09e52af15343909984359236a933ffe1f46ab7ada48929cc2741cd10782f\": rpc error: code = NotFound desc = could not find container \"52ca09e52af15343909984359236a933ffe1f46ab7ada48929cc2741cd10782f\": container with ID starting with 52ca09e52af15343909984359236a933ffe1f46ab7ada48929cc2741cd10782f not found: ID does not exist" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.353654 4714 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec0eba7e-2ea0-432f-bc57-d87404801abe-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.580751 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kg9qt"] Jan 29 16:13:54 crc kubenswrapper[4714]: I0129 16:13:54.586697 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kg9qt"] Jan 29 16:13:55 crc kubenswrapper[4714]: I0129 16:13:55.949005 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-h8b4r"] Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.189906 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="213a402c-b327-4aa6-9690-22d6da8664a4" path="/var/lib/kubelet/pods/213a402c-b327-4aa6-9690-22d6da8664a4/volumes" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.190692 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec0eba7e-2ea0-432f-bc57-d87404801abe" path="/var/lib/kubelet/pods/ec0eba7e-2ea0-432f-bc57-d87404801abe/volumes" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.494260 4714 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 29 16:13:56 crc kubenswrapper[4714]: E0129 16:13:56.495144 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec0eba7e-2ea0-432f-bc57-d87404801abe" containerName="extract-content" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.495167 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec0eba7e-2ea0-432f-bc57-d87404801abe" containerName="extract-content" Jan 29 16:13:56 crc kubenswrapper[4714]: E0129 16:13:56.495191 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="213a402c-b327-4aa6-9690-22d6da8664a4" containerName="extract-content" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.495198 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="213a402c-b327-4aa6-9690-22d6da8664a4" containerName="extract-content" Jan 29 16:13:56 crc kubenswrapper[4714]: E0129 16:13:56.495213 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef" containerName="extract-content" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.495224 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef" containerName="extract-content" Jan 29 16:13:56 crc kubenswrapper[4714]: E0129 16:13:56.495237 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef" containerName="extract-utilities" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.495244 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef" containerName="extract-utilities" Jan 29 16:13:56 crc kubenswrapper[4714]: E0129 16:13:56.495254 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb" containerName="pruner" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.495261 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb" containerName="pruner" Jan 29 16:13:56 crc kubenswrapper[4714]: E0129 16:13:56.495274 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec0eba7e-2ea0-432f-bc57-d87404801abe" containerName="registry-server" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.495283 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec0eba7e-2ea0-432f-bc57-d87404801abe" containerName="registry-server" Jan 29 16:13:56 crc kubenswrapper[4714]: E0129 16:13:56.495293 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef" containerName="registry-server" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.495299 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef" containerName="registry-server" Jan 29 16:13:56 crc kubenswrapper[4714]: E0129 16:13:56.495312 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="213a402c-b327-4aa6-9690-22d6da8664a4" containerName="registry-server" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.495318 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="213a402c-b327-4aa6-9690-22d6da8664a4" containerName="registry-server" Jan 29 16:13:56 crc kubenswrapper[4714]: E0129 16:13:56.495375 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="213a402c-b327-4aa6-9690-22d6da8664a4" containerName="extract-utilities" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.495382 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="213a402c-b327-4aa6-9690-22d6da8664a4" containerName="extract-utilities" Jan 29 16:13:56 crc kubenswrapper[4714]: E0129 16:13:56.495394 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec0eba7e-2ea0-432f-bc57-d87404801abe" containerName="extract-utilities" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.495400 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec0eba7e-2ea0-432f-bc57-d87404801abe" containerName="extract-utilities" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.496217 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae1263ae-fc9d-4d73-ac5d-65e9e4dacfef" containerName="registry-server" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.496239 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec0eba7e-2ea0-432f-bc57-d87404801abe" containerName="registry-server" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.496250 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="213a402c-b327-4aa6-9690-22d6da8664a4" containerName="registry-server" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.496268 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="c33e9e66-b87c-4a7a-8c7b-8fd44b4bd3eb" containerName="pruner" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.497337 4714 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.497731 4714 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.497798 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f" gracePeriod=15 Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.498047 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.498223 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79" gracePeriod=15 Jan 29 16:13:56 crc kubenswrapper[4714]: E0129 16:13:56.498346 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.498916 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 29 16:13:56 crc kubenswrapper[4714]: E0129 16:13:56.498948 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.498956 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 29 16:13:56 crc kubenswrapper[4714]: E0129 16:13:56.498988 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.499010 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 16:13:56 crc kubenswrapper[4714]: E0129 16:13:56.499026 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.499032 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 29 16:13:56 crc kubenswrapper[4714]: E0129 16:13:56.499063 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.499070 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 29 16:13:56 crc kubenswrapper[4714]: E0129 16:13:56.499084 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.499090 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.499286 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.499300 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.499307 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.499319 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.499347 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.499355 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 16:13:56 crc kubenswrapper[4714]: E0129 16:13:56.499584 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.499593 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.498381 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa" gracePeriod=15 Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.498416 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2" gracePeriod=15 Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.498480 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6" gracePeriod=15 Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.503748 4714 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.589658 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.589718 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.589745 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.589760 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.589777 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.589795 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.590019 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.590655 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.692145 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.692232 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.692269 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.692296 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.692323 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.692346 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.692366 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.692446 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.692535 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.692586 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.692613 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.692639 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.692668 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.692713 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.692765 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:13:56 crc kubenswrapper[4714]: I0129 16:13:56.692802 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:13:57 crc kubenswrapper[4714]: I0129 16:13:57.112445 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xtr82" Jan 29 16:13:57 crc kubenswrapper[4714]: I0129 16:13:57.113469 4714 status_manager.go:851] "Failed to get status for pod" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" pod="openshift-marketplace/community-operators-xtr82" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-xtr82\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:57 crc kubenswrapper[4714]: I0129 16:13:57.155127 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xtr82" Jan 29 16:13:57 crc kubenswrapper[4714]: I0129 16:13:57.155566 4714 status_manager.go:851] "Failed to get status for pod" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" pod="openshift-marketplace/community-operators-xtr82" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-xtr82\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:57 crc kubenswrapper[4714]: I0129 16:13:57.269577 4714 generic.go:334] "Generic (PLEG): container finished" podID="9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24" containerID="d7f3c5ef1cf90e64e41a56b19a6b215a0e1a7265f9a460a77da807a32640c2c8" exitCode=0 Jan 29 16:13:57 crc kubenswrapper[4714]: I0129 16:13:57.269679 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24","Type":"ContainerDied","Data":"d7f3c5ef1cf90e64e41a56b19a6b215a0e1a7265f9a460a77da807a32640c2c8"} Jan 29 16:13:57 crc kubenswrapper[4714]: I0129 16:13:57.270363 4714 status_manager.go:851] "Failed to get status for pod" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" pod="openshift-marketplace/community-operators-xtr82" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-xtr82\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:57 crc kubenswrapper[4714]: I0129 16:13:57.270733 4714 status_manager.go:851] "Failed to get status for pod" podUID="9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:57 crc kubenswrapper[4714]: I0129 16:13:57.272727 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 29 16:13:57 crc kubenswrapper[4714]: I0129 16:13:57.274069 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 29 16:13:57 crc kubenswrapper[4714]: I0129 16:13:57.275002 4714 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79" exitCode=0 Jan 29 16:13:57 crc kubenswrapper[4714]: I0129 16:13:57.275062 4714 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2" exitCode=0 Jan 29 16:13:57 crc kubenswrapper[4714]: I0129 16:13:57.275086 4714 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa" exitCode=0 Jan 29 16:13:57 crc kubenswrapper[4714]: I0129 16:13:57.275107 4714 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6" exitCode=2 Jan 29 16:13:57 crc kubenswrapper[4714]: I0129 16:13:57.275107 4714 scope.go:117] "RemoveContainer" containerID="6e458517de6b8ed3eb1e3e2b912a6d2134d1f4d5cada6affcb056fbe35defce8" Jan 29 16:13:57 crc kubenswrapper[4714]: I0129 16:13:57.569419 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6bjgq" Jan 29 16:13:57 crc kubenswrapper[4714]: I0129 16:13:57.570495 4714 status_manager.go:851] "Failed to get status for pod" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" pod="openshift-marketplace/community-operators-xtr82" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-xtr82\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:57 crc kubenswrapper[4714]: I0129 16:13:57.570653 4714 status_manager.go:851] "Failed to get status for pod" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" pod="openshift-marketplace/community-operators-6bjgq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6bjgq\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:57 crc kubenswrapper[4714]: I0129 16:13:57.570918 4714 status_manager.go:851] "Failed to get status for pod" podUID="9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.283825 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.573040 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.573523 4714 status_manager.go:851] "Failed to get status for pod" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" pod="openshift-marketplace/community-operators-xtr82" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-xtr82\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.573822 4714 status_manager.go:851] "Failed to get status for pod" podUID="9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.578735 4714 status_manager.go:851] "Failed to get status for pod" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" pod="openshift-marketplace/community-operators-6bjgq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6bjgq\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.653087 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24-kube-api-access\") pod \"9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24\" (UID: \"9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24\") " Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.653526 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24-var-lock\") pod \"9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24\" (UID: \"9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24\") " Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.653573 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24-kubelet-dir\") pod \"9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24\" (UID: \"9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24\") " Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.653637 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24-var-lock" (OuterVolumeSpecName: "var-lock") pod "9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24" (UID: "9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.653745 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24" (UID: "9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.654008 4714 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24-var-lock\") on node \"crc\" DevicePath \"\"" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.654024 4714 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.664265 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24" (UID: "9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.755872 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.853787 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.854431 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.854892 4714 status_manager.go:851] "Failed to get status for pod" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" pod="openshift-marketplace/community-operators-xtr82" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-xtr82\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.855086 4714 status_manager.go:851] "Failed to get status for pod" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" pod="openshift-marketplace/community-operators-6bjgq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6bjgq\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.855359 4714 status_manager.go:851] "Failed to get status for pod" podUID="9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.855811 4714 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.957970 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.958035 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.958091 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.958163 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.958163 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.958264 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.958577 4714 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.958602 4714 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 29 16:13:58 crc kubenswrapper[4714]: I0129 16:13:58.958614 4714 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 29 16:13:59 crc kubenswrapper[4714]: I0129 16:13:59.291538 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24","Type":"ContainerDied","Data":"c9979ec9683e16f0b3aadc58f285988aa9a20624d0e8ae1b706d16b4bb36c291"} Jan 29 16:13:59 crc kubenswrapper[4714]: I0129 16:13:59.291593 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9979ec9683e16f0b3aadc58f285988aa9a20624d0e8ae1b706d16b4bb36c291" Jan 29 16:13:59 crc kubenswrapper[4714]: I0129 16:13:59.291654 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 29 16:13:59 crc kubenswrapper[4714]: I0129 16:13:59.297222 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 29 16:13:59 crc kubenswrapper[4714]: I0129 16:13:59.300423 4714 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f" exitCode=0 Jan 29 16:13:59 crc kubenswrapper[4714]: I0129 16:13:59.300481 4714 scope.go:117] "RemoveContainer" containerID="422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79" Jan 29 16:13:59 crc kubenswrapper[4714]: I0129 16:13:59.300600 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:13:59 crc kubenswrapper[4714]: I0129 16:13:59.307434 4714 status_manager.go:851] "Failed to get status for pod" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" pod="openshift-marketplace/community-operators-xtr82" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-xtr82\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:59 crc kubenswrapper[4714]: I0129 16:13:59.308068 4714 status_manager.go:851] "Failed to get status for pod" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" pod="openshift-marketplace/community-operators-6bjgq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6bjgq\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:59 crc kubenswrapper[4714]: I0129 16:13:59.308392 4714 status_manager.go:851] "Failed to get status for pod" podUID="9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:59 crc kubenswrapper[4714]: I0129 16:13:59.308857 4714 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:59 crc kubenswrapper[4714]: I0129 16:13:59.313784 4714 status_manager.go:851] "Failed to get status for pod" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" pod="openshift-marketplace/community-operators-xtr82" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-xtr82\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:59 crc kubenswrapper[4714]: I0129 16:13:59.314209 4714 status_manager.go:851] "Failed to get status for pod" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" pod="openshift-marketplace/community-operators-6bjgq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6bjgq\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:59 crc kubenswrapper[4714]: I0129 16:13:59.314480 4714 status_manager.go:851] "Failed to get status for pod" podUID="9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:59 crc kubenswrapper[4714]: I0129 16:13:59.314824 4714 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:13:59 crc kubenswrapper[4714]: I0129 16:13:59.315535 4714 scope.go:117] "RemoveContainer" containerID="5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2" Jan 29 16:13:59 crc kubenswrapper[4714]: I0129 16:13:59.325283 4714 scope.go:117] "RemoveContainer" containerID="b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa" Jan 29 16:13:59 crc kubenswrapper[4714]: I0129 16:13:59.336027 4714 scope.go:117] "RemoveContainer" containerID="4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6" Jan 29 16:13:59 crc kubenswrapper[4714]: I0129 16:13:59.351165 4714 scope.go:117] "RemoveContainer" containerID="a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f" Jan 29 16:13:59 crc kubenswrapper[4714]: I0129 16:13:59.365491 4714 scope.go:117] "RemoveContainer" containerID="3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c" Jan 29 16:14:00 crc kubenswrapper[4714]: I0129 16:14:00.190907 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Jan 29 16:14:00 crc kubenswrapper[4714]: I0129 16:14:00.359925 4714 scope.go:117] "RemoveContainer" containerID="422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79" Jan 29 16:14:00 crc kubenswrapper[4714]: E0129 16:14:00.360621 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\": container with ID starting with 422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79 not found: ID does not exist" containerID="422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79" Jan 29 16:14:00 crc kubenswrapper[4714]: I0129 16:14:00.360666 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79"} err="failed to get container status \"422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\": rpc error: code = NotFound desc = could not find container \"422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79\": container with ID starting with 422e671f983882856957a72a899cd1e37d99ed4214cfd94bbf5b94933859ae79 not found: ID does not exist" Jan 29 16:14:00 crc kubenswrapper[4714]: I0129 16:14:00.360696 4714 scope.go:117] "RemoveContainer" containerID="5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2" Jan 29 16:14:00 crc kubenswrapper[4714]: E0129 16:14:00.361096 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\": container with ID starting with 5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2 not found: ID does not exist" containerID="5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2" Jan 29 16:14:00 crc kubenswrapper[4714]: I0129 16:14:00.361131 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2"} err="failed to get container status \"5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\": rpc error: code = NotFound desc = could not find container \"5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2\": container with ID starting with 5054d435e2cfce3893996ab2bd3746f95318408942e3239d3a184da25d540ba2 not found: ID does not exist" Jan 29 16:14:00 crc kubenswrapper[4714]: I0129 16:14:00.361158 4714 scope.go:117] "RemoveContainer" containerID="b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa" Jan 29 16:14:00 crc kubenswrapper[4714]: E0129 16:14:00.361486 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\": container with ID starting with b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa not found: ID does not exist" containerID="b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa" Jan 29 16:14:00 crc kubenswrapper[4714]: I0129 16:14:00.361526 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa"} err="failed to get container status \"b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\": rpc error: code = NotFound desc = could not find container \"b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa\": container with ID starting with b4b8556c6768e3aa3dba661bc9e6bcf3b159139eada076ff3419e54cbe0ca1fa not found: ID does not exist" Jan 29 16:14:00 crc kubenswrapper[4714]: I0129 16:14:00.361549 4714 scope.go:117] "RemoveContainer" containerID="4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6" Jan 29 16:14:00 crc kubenswrapper[4714]: E0129 16:14:00.361869 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\": container with ID starting with 4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6 not found: ID does not exist" containerID="4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6" Jan 29 16:14:00 crc kubenswrapper[4714]: I0129 16:14:00.361918 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6"} err="failed to get container status \"4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\": rpc error: code = NotFound desc = could not find container \"4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6\": container with ID starting with 4ab2f0e3ab0fcab15e30d6b29bb14662bcbb0c9357e4df00c15e43e3970193c6 not found: ID does not exist" Jan 29 16:14:00 crc kubenswrapper[4714]: I0129 16:14:00.361994 4714 scope.go:117] "RemoveContainer" containerID="a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f" Jan 29 16:14:00 crc kubenswrapper[4714]: E0129 16:14:00.364786 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\": container with ID starting with a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f not found: ID does not exist" containerID="a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f" Jan 29 16:14:00 crc kubenswrapper[4714]: I0129 16:14:00.364828 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f"} err="failed to get container status \"a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\": rpc error: code = NotFound desc = could not find container \"a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f\": container with ID starting with a754474ea4ff3c800aa715e042720449b861d7a7911e34ec627aa9a94807f30f not found: ID does not exist" Jan 29 16:14:00 crc kubenswrapper[4714]: I0129 16:14:00.364843 4714 scope.go:117] "RemoveContainer" containerID="3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c" Jan 29 16:14:00 crc kubenswrapper[4714]: E0129 16:14:00.365294 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\": container with ID starting with 3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c not found: ID does not exist" containerID="3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c" Jan 29 16:14:00 crc kubenswrapper[4714]: I0129 16:14:00.365329 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c"} err="failed to get container status \"3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\": rpc error: code = NotFound desc = could not find container \"3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c\": container with ID starting with 3af6d31edf201152cb07bcdbe09550306376ea701a5ab9d7755a552218c57c8c not found: ID does not exist" Jan 29 16:14:01 crc kubenswrapper[4714]: E0129 16:14:01.538677 4714 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.46:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:14:01 crc kubenswrapper[4714]: I0129 16:14:01.539173 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:14:01 crc kubenswrapper[4714]: W0129 16:14:01.558496 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-c2827e545ce1c2813ef6994a00eaf13c969a902919f03f1d776c0660aff8ea0d WatchSource:0}: Error finding container c2827e545ce1c2813ef6994a00eaf13c969a902919f03f1d776c0660aff8ea0d: Status 404 returned error can't find the container with id c2827e545ce1c2813ef6994a00eaf13c969a902919f03f1d776c0660aff8ea0d Jan 29 16:14:01 crc kubenswrapper[4714]: E0129 16:14:01.560644 4714 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.46:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188f3fb8cf10316f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-29 16:14:01.560371567 +0000 UTC m=+248.080872697,LastTimestamp:2026-01-29 16:14:01.560371567 +0000 UTC m=+248.080872697,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 29 16:14:02 crc kubenswrapper[4714]: I0129 16:14:02.323534 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"bab9b90f5d182c4ebe5eb811fac6ec3fd2ecd7c7c38b72ecfd4df4e4a0e90141"} Jan 29 16:14:02 crc kubenswrapper[4714]: I0129 16:14:02.323592 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"c2827e545ce1c2813ef6994a00eaf13c969a902919f03f1d776c0660aff8ea0d"} Jan 29 16:14:02 crc kubenswrapper[4714]: E0129 16:14:02.324233 4714 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.46:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:14:02 crc kubenswrapper[4714]: I0129 16:14:02.324484 4714 status_manager.go:851] "Failed to get status for pod" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" pod="openshift-marketplace/community-operators-xtr82" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-xtr82\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:02 crc kubenswrapper[4714]: I0129 16:14:02.324740 4714 status_manager.go:851] "Failed to get status for pod" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" pod="openshift-marketplace/community-operators-6bjgq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6bjgq\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:02 crc kubenswrapper[4714]: I0129 16:14:02.325076 4714 status_manager.go:851] "Failed to get status for pod" podUID="9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:04 crc kubenswrapper[4714]: I0129 16:14:04.189364 4714 status_manager.go:851] "Failed to get status for pod" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" pod="openshift-marketplace/community-operators-xtr82" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-xtr82\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:04 crc kubenswrapper[4714]: I0129 16:14:04.190301 4714 status_manager.go:851] "Failed to get status for pod" podUID="9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:04 crc kubenswrapper[4714]: I0129 16:14:04.190712 4714 status_manager.go:851] "Failed to get status for pod" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" pod="openshift-marketplace/community-operators-6bjgq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6bjgq\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:04 crc kubenswrapper[4714]: E0129 16:14:04.973787 4714 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:04 crc kubenswrapper[4714]: E0129 16:14:04.974775 4714 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:04 crc kubenswrapper[4714]: E0129 16:14:04.975471 4714 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:04 crc kubenswrapper[4714]: E0129 16:14:04.975904 4714 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:04 crc kubenswrapper[4714]: E0129 16:14:04.976493 4714 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:04 crc kubenswrapper[4714]: I0129 16:14:04.976543 4714 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Jan 29 16:14:04 crc kubenswrapper[4714]: E0129 16:14:04.976846 4714 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="200ms" Jan 29 16:14:05 crc kubenswrapper[4714]: E0129 16:14:05.178586 4714 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="400ms" Jan 29 16:14:05 crc kubenswrapper[4714]: E0129 16:14:05.579367 4714 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="800ms" Jan 29 16:14:06 crc kubenswrapper[4714]: E0129 16:14:06.380236 4714 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="1.6s" Jan 29 16:14:07 crc kubenswrapper[4714]: E0129 16:14:07.981248 4714 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="3.2s" Jan 29 16:14:08 crc kubenswrapper[4714]: I0129 16:14:08.183621 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:14:08 crc kubenswrapper[4714]: I0129 16:14:08.184560 4714 status_manager.go:851] "Failed to get status for pod" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" pod="openshift-marketplace/community-operators-xtr82" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-xtr82\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:08 crc kubenswrapper[4714]: I0129 16:14:08.185323 4714 status_manager.go:851] "Failed to get status for pod" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" pod="openshift-marketplace/community-operators-6bjgq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6bjgq\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:08 crc kubenswrapper[4714]: I0129 16:14:08.185810 4714 status_manager.go:851] "Failed to get status for pod" podUID="9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:08 crc kubenswrapper[4714]: I0129 16:14:08.208923 4714 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="810f3b92-c43d-41fd-8a5f-0f926ed63e50" Jan 29 16:14:08 crc kubenswrapper[4714]: I0129 16:14:08.208990 4714 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="810f3b92-c43d-41fd-8a5f-0f926ed63e50" Jan 29 16:14:08 crc kubenswrapper[4714]: E0129 16:14:08.209488 4714 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:14:08 crc kubenswrapper[4714]: I0129 16:14:08.210073 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:14:08 crc kubenswrapper[4714]: I0129 16:14:08.370844 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e76183f9859ceea5d6ba47762effc0bc5f31d2e9dcea477167d58699be798bdd"} Jan 29 16:14:09 crc kubenswrapper[4714]: I0129 16:14:09.378420 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 29 16:14:09 crc kubenswrapper[4714]: I0129 16:14:09.378457 4714 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835" exitCode=1 Jan 29 16:14:09 crc kubenswrapper[4714]: I0129 16:14:09.378502 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835"} Jan 29 16:14:09 crc kubenswrapper[4714]: I0129 16:14:09.378991 4714 scope.go:117] "RemoveContainer" containerID="d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835" Jan 29 16:14:09 crc kubenswrapper[4714]: I0129 16:14:09.381267 4714 status_manager.go:851] "Failed to get status for pod" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" pod="openshift-marketplace/community-operators-xtr82" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-xtr82\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:09 crc kubenswrapper[4714]: I0129 16:14:09.381663 4714 status_manager.go:851] "Failed to get status for pod" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" pod="openshift-marketplace/community-operators-6bjgq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6bjgq\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:09 crc kubenswrapper[4714]: I0129 16:14:09.382206 4714 status_manager.go:851] "Failed to get status for pod" podUID="9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:09 crc kubenswrapper[4714]: I0129 16:14:09.382491 4714 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="557886b8390c654504d86188f97a1e0330661ae0a5a81431ae900193973c2ff7" exitCode=0 Jan 29 16:14:09 crc kubenswrapper[4714]: I0129 16:14:09.382515 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"557886b8390c654504d86188f97a1e0330661ae0a5a81431ae900193973c2ff7"} Jan 29 16:14:09 crc kubenswrapper[4714]: I0129 16:14:09.382734 4714 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:09 crc kubenswrapper[4714]: I0129 16:14:09.382851 4714 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="810f3b92-c43d-41fd-8a5f-0f926ed63e50" Jan 29 16:14:09 crc kubenswrapper[4714]: I0129 16:14:09.382883 4714 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="810f3b92-c43d-41fd-8a5f-0f926ed63e50" Jan 29 16:14:09 crc kubenswrapper[4714]: E0129 16:14:09.383282 4714 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:14:09 crc kubenswrapper[4714]: I0129 16:14:09.383513 4714 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:09 crc kubenswrapper[4714]: I0129 16:14:09.383914 4714 status_manager.go:851] "Failed to get status for pod" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" pod="openshift-marketplace/community-operators-xtr82" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-xtr82\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:09 crc kubenswrapper[4714]: I0129 16:14:09.384357 4714 status_manager.go:851] "Failed to get status for pod" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" pod="openshift-marketplace/community-operators-6bjgq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6bjgq\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:09 crc kubenswrapper[4714]: I0129 16:14:09.385120 4714 status_manager.go:851] "Failed to get status for pod" podUID="9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Jan 29 16:14:10 crc kubenswrapper[4714]: I0129 16:14:10.177849 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:14:10 crc kubenswrapper[4714]: I0129 16:14:10.404612 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 29 16:14:10 crc kubenswrapper[4714]: I0129 16:14:10.404726 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e41d24707f09801e897cb0baf4db02b0644cf5ce6dabf14163675df4678cd1d1"} Jan 29 16:14:10 crc kubenswrapper[4714]: I0129 16:14:10.409670 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"62502093aedf338472b1261dc436b80ad6d8f7171817093db5cd839155c867a6"} Jan 29 16:14:10 crc kubenswrapper[4714]: I0129 16:14:10.409728 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4c7978fe28e5ff4501e1852333df37847b013416c0b696c89be556464d8091e5"} Jan 29 16:14:10 crc kubenswrapper[4714]: I0129 16:14:10.409747 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"47e1a9368e1ef86883e3c9c9f4b17768154fc3db5a8ae34a582625e9c46c2141"} Jan 29 16:14:11 crc kubenswrapper[4714]: I0129 16:14:11.416402 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"79e61e663a3f3661a37379cab961763d4262bf19a479e965afdef0e30b930772"} Jan 29 16:14:11 crc kubenswrapper[4714]: I0129 16:14:11.416451 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"908aa0992c2ec1ba47bc8f6aacd67caae83f412a5dc37bca91d4fa7c0135ee38"} Jan 29 16:14:11 crc kubenswrapper[4714]: I0129 16:14:11.416608 4714 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="810f3b92-c43d-41fd-8a5f-0f926ed63e50" Jan 29 16:14:11 crc kubenswrapper[4714]: I0129 16:14:11.416618 4714 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="810f3b92-c43d-41fd-8a5f-0f926ed63e50" Jan 29 16:14:13 crc kubenswrapper[4714]: I0129 16:14:13.210803 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:14:13 crc kubenswrapper[4714]: I0129 16:14:13.211257 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:14:13 crc kubenswrapper[4714]: I0129 16:14:13.220839 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:14:16 crc kubenswrapper[4714]: I0129 16:14:16.426216 4714 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:14:16 crc kubenswrapper[4714]: I0129 16:14:16.456919 4714 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"810f3b92-c43d-41fd-8a5f-0f926ed63e50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:14:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:14:09Z\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:14:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T16:14:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47e1a9368e1ef86883e3c9c9f4b17768154fc3db5a8ae34a582625e9c46c2141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:14:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62502093aedf338472b1261dc436b80ad6d8f7171817093db5cd839155c867a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:14:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c7978fe28e5ff4501e1852333df37847b013416c0b696c89be556464d8091e5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:14:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e61e663a3f3661a37379cab961763d4262bf19a479e965afdef0e30b930772\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:14:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://908aa0992c2ec1ba47bc8f6aacd67caae83f412a5dc37bca91d4fa7c0135ee38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T16:14:10Z\\\"}}}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://557886b8390c654504d86188f97a1e0330661ae0a5a81431ae900193973c2ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://557886b8390c654504d86188f97a1e0330661ae0a5a81431ae900193973c2ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T16:14:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T16:14:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}]}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Pod \"kube-apiserver-crc\" is invalid: metadata.uid: Invalid value: \"810f3b92-c43d-41fd-8a5f-0f926ed63e50\": field is immutable" Jan 29 16:14:17 crc kubenswrapper[4714]: I0129 16:14:17.450378 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:14:17 crc kubenswrapper[4714]: I0129 16:14:17.450510 4714 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="810f3b92-c43d-41fd-8a5f-0f926ed63e50" Jan 29 16:14:17 crc kubenswrapper[4714]: I0129 16:14:17.450547 4714 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="810f3b92-c43d-41fd-8a5f-0f926ed63e50" Jan 29 16:14:17 crc kubenswrapper[4714]: I0129 16:14:17.458689 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:14:17 crc kubenswrapper[4714]: I0129 16:14:17.464460 4714 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="d341acb7-4a33-42d9-b4bb-de8e3c376306" Jan 29 16:14:18 crc kubenswrapper[4714]: I0129 16:14:18.469017 4714 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="810f3b92-c43d-41fd-8a5f-0f926ed63e50" Jan 29 16:14:18 crc kubenswrapper[4714]: I0129 16:14:18.469339 4714 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="810f3b92-c43d-41fd-8a5f-0f926ed63e50" Jan 29 16:14:18 crc kubenswrapper[4714]: I0129 16:14:18.473798 4714 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="d341acb7-4a33-42d9-b4bb-de8e3c376306" Jan 29 16:14:18 crc kubenswrapper[4714]: I0129 16:14:18.513197 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:14:18 crc kubenswrapper[4714]: I0129 16:14:18.513767 4714 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 29 16:14:18 crc kubenswrapper[4714]: I0129 16:14:18.513918 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 29 16:14:19 crc kubenswrapper[4714]: I0129 16:14:19.481019 4714 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="810f3b92-c43d-41fd-8a5f-0f926ed63e50" Jan 29 16:14:19 crc kubenswrapper[4714]: I0129 16:14:19.481068 4714 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="810f3b92-c43d-41fd-8a5f-0f926ed63e50" Jan 29 16:14:19 crc kubenswrapper[4714]: I0129 16:14:19.485447 4714 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="d341acb7-4a33-42d9-b4bb-de8e3c376306" Jan 29 16:14:20 crc kubenswrapper[4714]: I0129 16:14:20.177669 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:14:20 crc kubenswrapper[4714]: I0129 16:14:20.978323 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" podUID="832097a5-4691-42b6-99cc-38679071d5ee" containerName="oauth-openshift" containerID="cri-o://9bc1cd1aee2de10059a78dc94ebfc1cb6a64c8e3be39806bf6e6ab107f47abff" gracePeriod=15 Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.452178 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.494052 4714 generic.go:334] "Generic (PLEG): container finished" podID="832097a5-4691-42b6-99cc-38679071d5ee" containerID="9bc1cd1aee2de10059a78dc94ebfc1cb6a64c8e3be39806bf6e6ab107f47abff" exitCode=0 Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.494102 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" event={"ID":"832097a5-4691-42b6-99cc-38679071d5ee","Type":"ContainerDied","Data":"9bc1cd1aee2de10059a78dc94ebfc1cb6a64c8e3be39806bf6e6ab107f47abff"} Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.494130 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" event={"ID":"832097a5-4691-42b6-99cc-38679071d5ee","Type":"ContainerDied","Data":"3a5ee9422c0e8f2bda4f13b1ec7a93ce78a161df42fc1dddfe6f8337aed30775"} Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.494133 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-h8b4r" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.494147 4714 scope.go:117] "RemoveContainer" containerID="9bc1cd1aee2de10059a78dc94ebfc1cb6a64c8e3be39806bf6e6ab107f47abff" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.512834 4714 scope.go:117] "RemoveContainer" containerID="9bc1cd1aee2de10059a78dc94ebfc1cb6a64c8e3be39806bf6e6ab107f47abff" Jan 29 16:14:21 crc kubenswrapper[4714]: E0129 16:14:21.513296 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bc1cd1aee2de10059a78dc94ebfc1cb6a64c8e3be39806bf6e6ab107f47abff\": container with ID starting with 9bc1cd1aee2de10059a78dc94ebfc1cb6a64c8e3be39806bf6e6ab107f47abff not found: ID does not exist" containerID="9bc1cd1aee2de10059a78dc94ebfc1cb6a64c8e3be39806bf6e6ab107f47abff" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.513343 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bc1cd1aee2de10059a78dc94ebfc1cb6a64c8e3be39806bf6e6ab107f47abff"} err="failed to get container status \"9bc1cd1aee2de10059a78dc94ebfc1cb6a64c8e3be39806bf6e6ab107f47abff\": rpc error: code = NotFound desc = could not find container \"9bc1cd1aee2de10059a78dc94ebfc1cb6a64c8e3be39806bf6e6ab107f47abff\": container with ID starting with 9bc1cd1aee2de10059a78dc94ebfc1cb6a64c8e3be39806bf6e6ab107f47abff not found: ID does not exist" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.604409 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-cliconfig\") pod \"832097a5-4691-42b6-99cc-38679071d5ee\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.604473 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-session\") pod \"832097a5-4691-42b6-99cc-38679071d5ee\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.604534 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-template-login\") pod \"832097a5-4691-42b6-99cc-38679071d5ee\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.604572 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8nmp\" (UniqueName: \"kubernetes.io/projected/832097a5-4691-42b6-99cc-38679071d5ee-kube-api-access-l8nmp\") pod \"832097a5-4691-42b6-99cc-38679071d5ee\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.604616 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-ocp-branding-template\") pod \"832097a5-4691-42b6-99cc-38679071d5ee\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.604654 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-router-certs\") pod \"832097a5-4691-42b6-99cc-38679071d5ee\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.604692 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-template-error\") pod \"832097a5-4691-42b6-99cc-38679071d5ee\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.604735 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/832097a5-4691-42b6-99cc-38679071d5ee-audit-dir\") pod \"832097a5-4691-42b6-99cc-38679071d5ee\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.604782 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-serving-cert\") pod \"832097a5-4691-42b6-99cc-38679071d5ee\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.604813 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-idp-0-file-data\") pod \"832097a5-4691-42b6-99cc-38679071d5ee\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.604863 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-audit-policies\") pod \"832097a5-4691-42b6-99cc-38679071d5ee\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.604903 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-service-ca\") pod \"832097a5-4691-42b6-99cc-38679071d5ee\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.604926 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/832097a5-4691-42b6-99cc-38679071d5ee-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "832097a5-4691-42b6-99cc-38679071d5ee" (UID: "832097a5-4691-42b6-99cc-38679071d5ee"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.604961 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-trusted-ca-bundle\") pod \"832097a5-4691-42b6-99cc-38679071d5ee\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.605061 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-template-provider-selection\") pod \"832097a5-4691-42b6-99cc-38679071d5ee\" (UID: \"832097a5-4691-42b6-99cc-38679071d5ee\") " Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.605390 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "832097a5-4691-42b6-99cc-38679071d5ee" (UID: "832097a5-4691-42b6-99cc-38679071d5ee"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.605582 4714 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/832097a5-4691-42b6-99cc-38679071d5ee-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.605612 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.605679 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "832097a5-4691-42b6-99cc-38679071d5ee" (UID: "832097a5-4691-42b6-99cc-38679071d5ee"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.606641 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "832097a5-4691-42b6-99cc-38679071d5ee" (UID: "832097a5-4691-42b6-99cc-38679071d5ee"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.606655 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "832097a5-4691-42b6-99cc-38679071d5ee" (UID: "832097a5-4691-42b6-99cc-38679071d5ee"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.610884 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "832097a5-4691-42b6-99cc-38679071d5ee" (UID: "832097a5-4691-42b6-99cc-38679071d5ee"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.610988 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/832097a5-4691-42b6-99cc-38679071d5ee-kube-api-access-l8nmp" (OuterVolumeSpecName: "kube-api-access-l8nmp") pod "832097a5-4691-42b6-99cc-38679071d5ee" (UID: "832097a5-4691-42b6-99cc-38679071d5ee"). InnerVolumeSpecName "kube-api-access-l8nmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.611409 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "832097a5-4691-42b6-99cc-38679071d5ee" (UID: "832097a5-4691-42b6-99cc-38679071d5ee"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.611720 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "832097a5-4691-42b6-99cc-38679071d5ee" (UID: "832097a5-4691-42b6-99cc-38679071d5ee"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.612022 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "832097a5-4691-42b6-99cc-38679071d5ee" (UID: "832097a5-4691-42b6-99cc-38679071d5ee"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.612216 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "832097a5-4691-42b6-99cc-38679071d5ee" (UID: "832097a5-4691-42b6-99cc-38679071d5ee"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.616236 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "832097a5-4691-42b6-99cc-38679071d5ee" (UID: "832097a5-4691-42b6-99cc-38679071d5ee"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.617395 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "832097a5-4691-42b6-99cc-38679071d5ee" (UID: "832097a5-4691-42b6-99cc-38679071d5ee"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.617526 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "832097a5-4691-42b6-99cc-38679071d5ee" (UID: "832097a5-4691-42b6-99cc-38679071d5ee"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.706671 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.706728 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.706752 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8nmp\" (UniqueName: \"kubernetes.io/projected/832097a5-4691-42b6-99cc-38679071d5ee-kube-api-access-l8nmp\") on node \"crc\" DevicePath \"\"" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.706773 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.706792 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.706811 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.706828 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.706846 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.706865 4714 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.706883 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.706900 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 16:14:21 crc kubenswrapper[4714]: I0129 16:14:21.706921 4714 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/832097a5-4691-42b6-99cc-38679071d5ee-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 29 16:14:25 crc kubenswrapper[4714]: I0129 16:14:25.924622 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 29 16:14:26 crc kubenswrapper[4714]: I0129 16:14:26.478571 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 29 16:14:26 crc kubenswrapper[4714]: I0129 16:14:26.879705 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 29 16:14:27 crc kubenswrapper[4714]: I0129 16:14:27.360789 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 29 16:14:27 crc kubenswrapper[4714]: I0129 16:14:27.472355 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 29 16:14:27 crc kubenswrapper[4714]: I0129 16:14:27.565722 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 29 16:14:27 crc kubenswrapper[4714]: I0129 16:14:27.649273 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 29 16:14:27 crc kubenswrapper[4714]: I0129 16:14:27.751321 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 29 16:14:27 crc kubenswrapper[4714]: I0129 16:14:27.802050 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 29 16:14:27 crc kubenswrapper[4714]: I0129 16:14:27.869251 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 29 16:14:28 crc kubenswrapper[4714]: I0129 16:14:28.222259 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 29 16:14:28 crc kubenswrapper[4714]: I0129 16:14:28.513413 4714 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 29 16:14:28 crc kubenswrapper[4714]: I0129 16:14:28.513488 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 29 16:14:28 crc kubenswrapper[4714]: I0129 16:14:28.585985 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 29 16:14:28 crc kubenswrapper[4714]: I0129 16:14:28.696317 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 29 16:14:28 crc kubenswrapper[4714]: I0129 16:14:28.788207 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 29 16:14:28 crc kubenswrapper[4714]: I0129 16:14:28.939879 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 29 16:14:29 crc kubenswrapper[4714]: I0129 16:14:29.348190 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 29 16:14:29 crc kubenswrapper[4714]: I0129 16:14:29.389566 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 29 16:14:30 crc kubenswrapper[4714]: I0129 16:14:30.015694 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 29 16:14:30 crc kubenswrapper[4714]: I0129 16:14:30.326636 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 29 16:14:30 crc kubenswrapper[4714]: I0129 16:14:30.335299 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 29 16:14:30 crc kubenswrapper[4714]: I0129 16:14:30.481724 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 29 16:14:30 crc kubenswrapper[4714]: I0129 16:14:30.502574 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 29 16:14:30 crc kubenswrapper[4714]: I0129 16:14:30.517517 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 29 16:14:30 crc kubenswrapper[4714]: I0129 16:14:30.654174 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 29 16:14:30 crc kubenswrapper[4714]: I0129 16:14:30.676918 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 29 16:14:30 crc kubenswrapper[4714]: I0129 16:14:30.697743 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 29 16:14:30 crc kubenswrapper[4714]: I0129 16:14:30.733763 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 29 16:14:30 crc kubenswrapper[4714]: I0129 16:14:30.860426 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 29 16:14:30 crc kubenswrapper[4714]: I0129 16:14:30.986107 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.038215 4714 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.058012 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.085042 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.176563 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.186671 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.241506 4714 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.297161 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.387720 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.413373 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.433537 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.488522 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.490775 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.504069 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.580899 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.586343 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.649732 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.675018 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.716404 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.765726 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.775541 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 29 16:14:31 crc kubenswrapper[4714]: I0129 16:14:31.790558 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 29 16:14:32 crc kubenswrapper[4714]: I0129 16:14:32.000871 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 29 16:14:32 crc kubenswrapper[4714]: I0129 16:14:32.048986 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 29 16:14:32 crc kubenswrapper[4714]: I0129 16:14:32.118777 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 29 16:14:32 crc kubenswrapper[4714]: I0129 16:14:32.123501 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 29 16:14:32 crc kubenswrapper[4714]: I0129 16:14:32.249262 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 29 16:14:32 crc kubenswrapper[4714]: I0129 16:14:32.249262 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 29 16:14:32 crc kubenswrapper[4714]: I0129 16:14:32.332242 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 29 16:14:32 crc kubenswrapper[4714]: I0129 16:14:32.354099 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 29 16:14:32 crc kubenswrapper[4714]: I0129 16:14:32.373598 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 29 16:14:32 crc kubenswrapper[4714]: I0129 16:14:32.500759 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 29 16:14:32 crc kubenswrapper[4714]: I0129 16:14:32.589645 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 29 16:14:32 crc kubenswrapper[4714]: I0129 16:14:32.673304 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 29 16:14:32 crc kubenswrapper[4714]: I0129 16:14:32.699695 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 29 16:14:32 crc kubenswrapper[4714]: I0129 16:14:32.721216 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 29 16:14:32 crc kubenswrapper[4714]: I0129 16:14:32.835761 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 29 16:14:32 crc kubenswrapper[4714]: I0129 16:14:32.959689 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 29 16:14:32 crc kubenswrapper[4714]: I0129 16:14:32.962276 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 29 16:14:32 crc kubenswrapper[4714]: I0129 16:14:32.978901 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 29 16:14:33 crc kubenswrapper[4714]: I0129 16:14:33.080383 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 29 16:14:33 crc kubenswrapper[4714]: I0129 16:14:33.179509 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 29 16:14:33 crc kubenswrapper[4714]: I0129 16:14:33.285836 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 29 16:14:33 crc kubenswrapper[4714]: I0129 16:14:33.293658 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 29 16:14:33 crc kubenswrapper[4714]: I0129 16:14:33.392287 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 29 16:14:33 crc kubenswrapper[4714]: I0129 16:14:33.410685 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 29 16:14:33 crc kubenswrapper[4714]: I0129 16:14:33.473831 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 29 16:14:33 crc kubenswrapper[4714]: I0129 16:14:33.672894 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 29 16:14:33 crc kubenswrapper[4714]: I0129 16:14:33.685036 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 29 16:14:33 crc kubenswrapper[4714]: I0129 16:14:33.789210 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 29 16:14:33 crc kubenswrapper[4714]: I0129 16:14:33.844323 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 29 16:14:33 crc kubenswrapper[4714]: I0129 16:14:33.878921 4714 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 29 16:14:33 crc kubenswrapper[4714]: I0129 16:14:33.887408 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 29 16:14:33 crc kubenswrapper[4714]: I0129 16:14:33.889633 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 29 16:14:33 crc kubenswrapper[4714]: I0129 16:14:33.948954 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 29 16:14:34 crc kubenswrapper[4714]: I0129 16:14:34.108735 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 29 16:14:34 crc kubenswrapper[4714]: I0129 16:14:34.239762 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 29 16:14:34 crc kubenswrapper[4714]: I0129 16:14:34.333830 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 29 16:14:34 crc kubenswrapper[4714]: I0129 16:14:34.383057 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 29 16:14:34 crc kubenswrapper[4714]: I0129 16:14:34.398480 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 29 16:14:34 crc kubenswrapper[4714]: I0129 16:14:34.547737 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 29 16:14:34 crc kubenswrapper[4714]: I0129 16:14:34.556730 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 29 16:14:34 crc kubenswrapper[4714]: I0129 16:14:34.621641 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 29 16:14:34 crc kubenswrapper[4714]: I0129 16:14:34.679257 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 29 16:14:34 crc kubenswrapper[4714]: I0129 16:14:34.681802 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 29 16:14:34 crc kubenswrapper[4714]: I0129 16:14:34.701506 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 29 16:14:34 crc kubenswrapper[4714]: I0129 16:14:34.723289 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 29 16:14:34 crc kubenswrapper[4714]: I0129 16:14:34.766008 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 29 16:14:34 crc kubenswrapper[4714]: I0129 16:14:34.935035 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 29 16:14:35 crc kubenswrapper[4714]: I0129 16:14:35.082392 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 29 16:14:35 crc kubenswrapper[4714]: I0129 16:14:35.121414 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 29 16:14:35 crc kubenswrapper[4714]: I0129 16:14:35.216967 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 29 16:14:35 crc kubenswrapper[4714]: I0129 16:14:35.292146 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 29 16:14:35 crc kubenswrapper[4714]: I0129 16:14:35.329058 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 29 16:14:35 crc kubenswrapper[4714]: I0129 16:14:35.329057 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 29 16:14:35 crc kubenswrapper[4714]: I0129 16:14:35.374417 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 29 16:14:35 crc kubenswrapper[4714]: I0129 16:14:35.426806 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 29 16:14:35 crc kubenswrapper[4714]: I0129 16:14:35.481877 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 29 16:14:35 crc kubenswrapper[4714]: I0129 16:14:35.546432 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 29 16:14:35 crc kubenswrapper[4714]: I0129 16:14:35.616915 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 29 16:14:35 crc kubenswrapper[4714]: I0129 16:14:35.662851 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 29 16:14:35 crc kubenswrapper[4714]: I0129 16:14:35.734541 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 29 16:14:35 crc kubenswrapper[4714]: I0129 16:14:35.742604 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 29 16:14:35 crc kubenswrapper[4714]: I0129 16:14:35.833239 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 29 16:14:35 crc kubenswrapper[4714]: I0129 16:14:35.845876 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 29 16:14:35 crc kubenswrapper[4714]: I0129 16:14:35.875121 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 29 16:14:35 crc kubenswrapper[4714]: I0129 16:14:35.902661 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 29 16:14:36 crc kubenswrapper[4714]: I0129 16:14:36.089869 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 29 16:14:36 crc kubenswrapper[4714]: I0129 16:14:36.177224 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 29 16:14:36 crc kubenswrapper[4714]: I0129 16:14:36.277910 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 29 16:14:36 crc kubenswrapper[4714]: I0129 16:14:36.343140 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 29 16:14:36 crc kubenswrapper[4714]: I0129 16:14:36.369839 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 29 16:14:36 crc kubenswrapper[4714]: I0129 16:14:36.379487 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 29 16:14:36 crc kubenswrapper[4714]: I0129 16:14:36.408357 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 29 16:14:36 crc kubenswrapper[4714]: I0129 16:14:36.512516 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 29 16:14:36 crc kubenswrapper[4714]: I0129 16:14:36.544443 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 29 16:14:36 crc kubenswrapper[4714]: I0129 16:14:36.650491 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 29 16:14:36 crc kubenswrapper[4714]: I0129 16:14:36.667915 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 29 16:14:36 crc kubenswrapper[4714]: I0129 16:14:36.669805 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 29 16:14:36 crc kubenswrapper[4714]: I0129 16:14:36.692864 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 29 16:14:36 crc kubenswrapper[4714]: I0129 16:14:36.757064 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 29 16:14:36 crc kubenswrapper[4714]: I0129 16:14:36.780365 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 29 16:14:36 crc kubenswrapper[4714]: I0129 16:14:36.842593 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 29 16:14:36 crc kubenswrapper[4714]: I0129 16:14:36.868861 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 29 16:14:36 crc kubenswrapper[4714]: I0129 16:14:36.981490 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 29 16:14:36 crc kubenswrapper[4714]: I0129 16:14:36.995670 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.025002 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.111419 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.205792 4714 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.210901 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-h8b4r"] Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.210990 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg","openshift-kube-apiserver/kube-apiserver-crc"] Jan 29 16:14:37 crc kubenswrapper[4714]: E0129 16:14:37.211182 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24" containerName="installer" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.211195 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24" containerName="installer" Jan 29 16:14:37 crc kubenswrapper[4714]: E0129 16:14:37.211207 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="832097a5-4691-42b6-99cc-38679071d5ee" containerName="oauth-openshift" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.211215 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="832097a5-4691-42b6-99cc-38679071d5ee" containerName="oauth-openshift" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.211333 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="832097a5-4691-42b6-99cc-38679071d5ee" containerName="oauth-openshift" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.211348 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cd159bf-dc15-4cdb-a61a-97c8b5ee4f24" containerName="installer" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.211705 4714 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="810f3b92-c43d-41fd-8a5f-0f926ed63e50" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.211745 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.211754 4714 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="810f3b92-c43d-41fd-8a5f-0f926ed63e50" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.213073 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.214632 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.215494 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.215692 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.215846 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.216091 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.216154 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.216525 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.219475 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.222705 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.222916 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.223130 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.223144 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.224060 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.244480 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.255725 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.260739 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.273008 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.280663 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.287665 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=21.287643797 podStartE2EDuration="21.287643797s" podCreationTimestamp="2026-01-29 16:14:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:14:37.286552176 +0000 UTC m=+283.807053336" watchObservedRunningTime="2026-01-29 16:14:37.287643797 +0000 UTC m=+283.808144957" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.332153 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.347464 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.347608 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.347655 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4464af5b-c16f-46b7-b29b-c22fa78843f3-audit-dir\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.347686 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-user-template-login\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.347737 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-service-ca\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.347832 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.347868 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-session\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.347894 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-user-template-error\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.347929 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.348001 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.348113 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-router-certs\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.348195 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fdhw\" (UniqueName: \"kubernetes.io/projected/4464af5b-c16f-46b7-b29b-c22fa78843f3-kube-api-access-6fdhw\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.348239 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4464af5b-c16f-46b7-b29b-c22fa78843f3-audit-policies\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.348292 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.377726 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.389041 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.412304 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.431630 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.437049 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.449778 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fdhw\" (UniqueName: \"kubernetes.io/projected/4464af5b-c16f-46b7-b29b-c22fa78843f3-kube-api-access-6fdhw\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.449866 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4464af5b-c16f-46b7-b29b-c22fa78843f3-audit-policies\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.450012 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.450092 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.450156 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.450218 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4464af5b-c16f-46b7-b29b-c22fa78843f3-audit-dir\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.450268 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-user-template-login\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.450351 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-service-ca\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.450420 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.450449 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4464af5b-c16f-46b7-b29b-c22fa78843f3-audit-dir\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.450477 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-session\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.451517 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4464af5b-c16f-46b7-b29b-c22fa78843f3-audit-policies\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.451761 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-user-template-error\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.451834 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.452308 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.452407 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-router-certs\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.452412 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.454375 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.456464 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.456557 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-service-ca\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.456735 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-user-template-login\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.457571 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.460117 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.465631 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.466051 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-user-template-error\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.467273 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-session\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.469483 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4464af5b-c16f-46b7-b29b-c22fa78843f3-v4-0-config-system-router-certs\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.481549 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fdhw\" (UniqueName: \"kubernetes.io/projected/4464af5b-c16f-46b7-b29b-c22fa78843f3-kube-api-access-6fdhw\") pod \"oauth-openshift-7cddd88c7f-kq5gg\" (UID: \"4464af5b-c16f-46b7-b29b-c22fa78843f3\") " pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.538841 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.581228 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.647691 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.648035 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.669696 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.709029 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.711407 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.735171 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.827460 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg"] Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.978674 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 29 16:14:37 crc kubenswrapper[4714]: I0129 16:14:37.988638 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.098864 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.131277 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.164263 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.166391 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.195651 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="832097a5-4691-42b6-99cc-38679071d5ee" path="/var/lib/kubelet/pods/832097a5-4691-42b6-99cc-38679071d5ee/volumes" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.241272 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.274486 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.293597 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.363635 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.467989 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.513595 4714 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.513972 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.514042 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.514974 4714 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"e41d24707f09801e897cb0baf4db02b0644cf5ce6dabf14163675df4678cd1d1"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.515168 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://e41d24707f09801e897cb0baf4db02b0644cf5ce6dabf14163675df4678cd1d1" gracePeriod=30 Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.533600 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.574999 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.592806 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.593744 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.635460 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.649716 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" event={"ID":"4464af5b-c16f-46b7-b29b-c22fa78843f3","Type":"ContainerStarted","Data":"fa827496154097254ba766d8a3902e298682ee4add2b5c9ea235ac1d29c3d416"} Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.649793 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" event={"ID":"4464af5b-c16f-46b7-b29b-c22fa78843f3","Type":"ContainerStarted","Data":"8160a991a75a27207a1d80dd0e4fdf5282d45ed57f5710eb653b00bc218edd8f"} Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.650361 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.693370 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.818374 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.837949 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.901967 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.908429 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.935233 4714 patch_prober.go:28] interesting pod/oauth-openshift-7cddd88c7f-kq5gg container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.56:6443/healthz\": read tcp 10.217.0.2:49984->10.217.0.56:6443: read: connection reset by peer" start-of-body= Jan 29 16:14:38 crc kubenswrapper[4714]: I0129 16:14:38.935296 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" podUID="4464af5b-c16f-46b7-b29b-c22fa78843f3" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.56:6443/healthz\": read tcp 10.217.0.2:49984->10.217.0.56:6443: read: connection reset by peer" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.001566 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.028143 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.050531 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.061248 4714 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.061586 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://bab9b90f5d182c4ebe5eb811fac6ec3fd2ecd7c7c38b72ecfd4df4e4a0e90141" gracePeriod=5 Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.100113 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.103401 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.106042 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.122825 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.135882 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.205587 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.211091 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.218550 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.322473 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.335116 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.363275 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.402051 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.411506 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.479644 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.486811 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.657181 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-7cddd88c7f-kq5gg_4464af5b-c16f-46b7-b29b-c22fa78843f3/oauth-openshift/0.log" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.657230 4714 generic.go:334] "Generic (PLEG): container finished" podID="4464af5b-c16f-46b7-b29b-c22fa78843f3" containerID="fa827496154097254ba766d8a3902e298682ee4add2b5c9ea235ac1d29c3d416" exitCode=255 Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.657260 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" event={"ID":"4464af5b-c16f-46b7-b29b-c22fa78843f3","Type":"ContainerDied","Data":"fa827496154097254ba766d8a3902e298682ee4add2b5c9ea235ac1d29c3d416"} Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.657704 4714 scope.go:117] "RemoveContainer" containerID="fa827496154097254ba766d8a3902e298682ee4add2b5c9ea235ac1d29c3d416" Jan 29 16:14:39 crc kubenswrapper[4714]: I0129 16:14:39.957246 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 29 16:14:40 crc kubenswrapper[4714]: I0129 16:14:40.017594 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 29 16:14:40 crc kubenswrapper[4714]: I0129 16:14:40.056060 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 29 16:14:40 crc kubenswrapper[4714]: I0129 16:14:40.074373 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 29 16:14:40 crc kubenswrapper[4714]: I0129 16:14:40.141630 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 29 16:14:40 crc kubenswrapper[4714]: I0129 16:14:40.164994 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 29 16:14:40 crc kubenswrapper[4714]: I0129 16:14:40.205080 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 29 16:14:40 crc kubenswrapper[4714]: I0129 16:14:40.305734 4714 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 29 16:14:40 crc kubenswrapper[4714]: I0129 16:14:40.364623 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 29 16:14:40 crc kubenswrapper[4714]: I0129 16:14:40.576116 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 29 16:14:40 crc kubenswrapper[4714]: I0129 16:14:40.665974 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-7cddd88c7f-kq5gg_4464af5b-c16f-46b7-b29b-c22fa78843f3/oauth-openshift/0.log" Jan 29 16:14:40 crc kubenswrapper[4714]: I0129 16:14:40.666036 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" event={"ID":"4464af5b-c16f-46b7-b29b-c22fa78843f3","Type":"ContainerStarted","Data":"9abc3d29fc5ca7a4d8cf8aa61dd76ea3277aa26f99a7eaae8b274e765a003eda"} Jan 29 16:14:40 crc kubenswrapper[4714]: I0129 16:14:40.667431 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:40 crc kubenswrapper[4714]: I0129 16:14:40.676674 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" Jan 29 16:14:40 crc kubenswrapper[4714]: I0129 16:14:40.697832 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7cddd88c7f-kq5gg" podStartSLOduration=45.697816919 podStartE2EDuration="45.697816919s" podCreationTimestamp="2026-01-29 16:13:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:14:38.681828364 +0000 UTC m=+285.202329514" watchObservedRunningTime="2026-01-29 16:14:40.697816919 +0000 UTC m=+287.218318039" Jan 29 16:14:40 crc kubenswrapper[4714]: I0129 16:14:40.843658 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 29 16:14:40 crc kubenswrapper[4714]: I0129 16:14:40.876437 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 29 16:14:40 crc kubenswrapper[4714]: I0129 16:14:40.889398 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 29 16:14:40 crc kubenswrapper[4714]: I0129 16:14:40.998437 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 29 16:14:41 crc kubenswrapper[4714]: I0129 16:14:41.047591 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 29 16:14:41 crc kubenswrapper[4714]: I0129 16:14:41.072760 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 29 16:14:41 crc kubenswrapper[4714]: I0129 16:14:41.170079 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 29 16:14:41 crc kubenswrapper[4714]: I0129 16:14:41.266330 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 29 16:14:41 crc kubenswrapper[4714]: I0129 16:14:41.325698 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 29 16:14:41 crc kubenswrapper[4714]: I0129 16:14:41.402810 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 29 16:14:41 crc kubenswrapper[4714]: I0129 16:14:41.543523 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 29 16:14:41 crc kubenswrapper[4714]: I0129 16:14:41.548628 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 29 16:14:41 crc kubenswrapper[4714]: I0129 16:14:41.587522 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 29 16:14:42 crc kubenswrapper[4714]: I0129 16:14:42.038985 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 29 16:14:42 crc kubenswrapper[4714]: I0129 16:14:42.115274 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 29 16:14:42 crc kubenswrapper[4714]: I0129 16:14:42.181925 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 29 16:14:42 crc kubenswrapper[4714]: I0129 16:14:42.226878 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 29 16:14:42 crc kubenswrapper[4714]: I0129 16:14:42.245580 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 29 16:14:42 crc kubenswrapper[4714]: I0129 16:14:42.301654 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 29 16:14:42 crc kubenswrapper[4714]: I0129 16:14:42.376251 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 29 16:14:42 crc kubenswrapper[4714]: I0129 16:14:42.497661 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 29 16:14:42 crc kubenswrapper[4714]: I0129 16:14:42.735457 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 29 16:14:42 crc kubenswrapper[4714]: I0129 16:14:42.823230 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 29 16:14:42 crc kubenswrapper[4714]: I0129 16:14:42.973979 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 29 16:14:43 crc kubenswrapper[4714]: I0129 16:14:43.087247 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 29 16:14:43 crc kubenswrapper[4714]: I0129 16:14:43.182886 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 29 16:14:43 crc kubenswrapper[4714]: I0129 16:14:43.325490 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 29 16:14:43 crc kubenswrapper[4714]: I0129 16:14:43.347348 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 29 16:14:43 crc kubenswrapper[4714]: I0129 16:14:43.988411 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.112562 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.128479 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.342895 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.607391 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.621384 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.667130 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.667258 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.691522 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.691598 4714 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="bab9b90f5d182c4ebe5eb811fac6ec3fd2ecd7c7c38b72ecfd4df4e4a0e90141" exitCode=137 Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.691638 4714 scope.go:117] "RemoveContainer" containerID="bab9b90f5d182c4ebe5eb811fac6ec3fd2ecd7c7c38b72ecfd4df4e4a0e90141" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.691741 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.716562 4714 scope.go:117] "RemoveContainer" containerID="bab9b90f5d182c4ebe5eb811fac6ec3fd2ecd7c7c38b72ecfd4df4e4a0e90141" Jan 29 16:14:44 crc kubenswrapper[4714]: E0129 16:14:44.717175 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bab9b90f5d182c4ebe5eb811fac6ec3fd2ecd7c7c38b72ecfd4df4e4a0e90141\": container with ID starting with bab9b90f5d182c4ebe5eb811fac6ec3fd2ecd7c7c38b72ecfd4df4e4a0e90141 not found: ID does not exist" containerID="bab9b90f5d182c4ebe5eb811fac6ec3fd2ecd7c7c38b72ecfd4df4e4a0e90141" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.717220 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bab9b90f5d182c4ebe5eb811fac6ec3fd2ecd7c7c38b72ecfd4df4e4a0e90141"} err="failed to get container status \"bab9b90f5d182c4ebe5eb811fac6ec3fd2ecd7c7c38b72ecfd4df4e4a0e90141\": rpc error: code = NotFound desc = could not find container \"bab9b90f5d182c4ebe5eb811fac6ec3fd2ecd7c7c38b72ecfd4df4e4a0e90141\": container with ID starting with bab9b90f5d182c4ebe5eb811fac6ec3fd2ecd7c7c38b72ecfd4df4e4a0e90141 not found: ID does not exist" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.853309 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.853413 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.853461 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.853545 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.853576 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.853639 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.853656 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.853698 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.853817 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.854062 4714 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.854098 4714 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.854126 4714 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.854150 4714 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.866177 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:14:44 crc kubenswrapper[4714]: I0129 16:14:44.956945 4714 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 29 16:14:45 crc kubenswrapper[4714]: I0129 16:14:45.373359 4714 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 29 16:14:46 crc kubenswrapper[4714]: I0129 16:14:46.195132 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Jan 29 16:14:53 crc kubenswrapper[4714]: I0129 16:14:53.987133 4714 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Jan 29 16:15:08 crc kubenswrapper[4714]: I0129 16:15:08.851220 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Jan 29 16:15:08 crc kubenswrapper[4714]: I0129 16:15:08.856782 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 29 16:15:08 crc kubenswrapper[4714]: I0129 16:15:08.856874 4714 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="e41d24707f09801e897cb0baf4db02b0644cf5ce6dabf14163675df4678cd1d1" exitCode=137 Jan 29 16:15:08 crc kubenswrapper[4714]: I0129 16:15:08.856920 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"e41d24707f09801e897cb0baf4db02b0644cf5ce6dabf14163675df4678cd1d1"} Jan 29 16:15:08 crc kubenswrapper[4714]: I0129 16:15:08.857023 4714 scope.go:117] "RemoveContainer" containerID="d8783a80e27263947194a6eff395452265c090e1a3e8155cb7cf12ec80eea835" Jan 29 16:15:09 crc kubenswrapper[4714]: I0129 16:15:09.866698 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Jan 29 16:15:09 crc kubenswrapper[4714]: I0129 16:15:09.868775 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a65c6860cf9526e7c6f9b8c6dca74bc104279387d0f33f88b0361f6e16f3cd32"} Jan 29 16:15:10 crc kubenswrapper[4714]: I0129 16:15:10.178461 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:15:18 crc kubenswrapper[4714]: I0129 16:15:18.513706 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:15:18 crc kubenswrapper[4714]: I0129 16:15:18.521547 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:15:20 crc kubenswrapper[4714]: I0129 16:15:20.195276 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.073900 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29495055-4hxfc"] Jan 29 16:15:28 crc kubenswrapper[4714]: E0129 16:15:28.074500 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.074511 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.074595 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.074953 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495055-4hxfc" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.076714 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.081211 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29495055-4hxfc"] Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.083026 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.102145 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjwmg\" (UniqueName: \"kubernetes.io/projected/d0d972c0-6998-401f-8f0a-5bea6ed5590f-kube-api-access-sjwmg\") pod \"collect-profiles-29495055-4hxfc\" (UID: \"d0d972c0-6998-401f-8f0a-5bea6ed5590f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495055-4hxfc" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.102187 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0d972c0-6998-401f-8f0a-5bea6ed5590f-secret-volume\") pod \"collect-profiles-29495055-4hxfc\" (UID: \"d0d972c0-6998-401f-8f0a-5bea6ed5590f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495055-4hxfc" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.102234 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0d972c0-6998-401f-8f0a-5bea6ed5590f-config-volume\") pod \"collect-profiles-29495055-4hxfc\" (UID: \"d0d972c0-6998-401f-8f0a-5bea6ed5590f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495055-4hxfc" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.109375 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xlczd"] Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.109586 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" podUID="3c2d0611-58f8-4a7e-8280-361c80d62802" containerName="controller-manager" containerID="cri-o://1a48e7995b6652b74e320bf25807bcc51cd5f49496615ac9028d0bf40f37019e" gracePeriod=30 Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.115829 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw"] Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.116031 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" podUID="fbfdd647-1d64-4d35-9af2-6dee52b4c860" containerName="route-controller-manager" containerID="cri-o://3e427856d41457853f68d5a32dc2f87168c0f52da592bc0a2b3db34f710fea2b" gracePeriod=30 Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.203048 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0d972c0-6998-401f-8f0a-5bea6ed5590f-config-volume\") pod \"collect-profiles-29495055-4hxfc\" (UID: \"d0d972c0-6998-401f-8f0a-5bea6ed5590f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495055-4hxfc" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.203119 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjwmg\" (UniqueName: \"kubernetes.io/projected/d0d972c0-6998-401f-8f0a-5bea6ed5590f-kube-api-access-sjwmg\") pod \"collect-profiles-29495055-4hxfc\" (UID: \"d0d972c0-6998-401f-8f0a-5bea6ed5590f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495055-4hxfc" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.203146 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0d972c0-6998-401f-8f0a-5bea6ed5590f-secret-volume\") pod \"collect-profiles-29495055-4hxfc\" (UID: \"d0d972c0-6998-401f-8f0a-5bea6ed5590f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495055-4hxfc" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.204188 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0d972c0-6998-401f-8f0a-5bea6ed5590f-config-volume\") pod \"collect-profiles-29495055-4hxfc\" (UID: \"d0d972c0-6998-401f-8f0a-5bea6ed5590f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495055-4hxfc" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.209692 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0d972c0-6998-401f-8f0a-5bea6ed5590f-secret-volume\") pod \"collect-profiles-29495055-4hxfc\" (UID: \"d0d972c0-6998-401f-8f0a-5bea6ed5590f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495055-4hxfc" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.233994 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjwmg\" (UniqueName: \"kubernetes.io/projected/d0d972c0-6998-401f-8f0a-5bea6ed5590f-kube-api-access-sjwmg\") pod \"collect-profiles-29495055-4hxfc\" (UID: \"d0d972c0-6998-401f-8f0a-5bea6ed5590f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495055-4hxfc" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.390246 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495055-4hxfc" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.490864 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.511872 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fbfdd647-1d64-4d35-9af2-6dee52b4c860-serving-cert\") pod \"fbfdd647-1d64-4d35-9af2-6dee52b4c860\" (UID: \"fbfdd647-1d64-4d35-9af2-6dee52b4c860\") " Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.511909 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbfdd647-1d64-4d35-9af2-6dee52b4c860-config\") pod \"fbfdd647-1d64-4d35-9af2-6dee52b4c860\" (UID: \"fbfdd647-1d64-4d35-9af2-6dee52b4c860\") " Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.511946 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fbfdd647-1d64-4d35-9af2-6dee52b4c860-client-ca\") pod \"fbfdd647-1d64-4d35-9af2-6dee52b4c860\" (UID: \"fbfdd647-1d64-4d35-9af2-6dee52b4c860\") " Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.512016 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srhd7\" (UniqueName: \"kubernetes.io/projected/fbfdd647-1d64-4d35-9af2-6dee52b4c860-kube-api-access-srhd7\") pod \"fbfdd647-1d64-4d35-9af2-6dee52b4c860\" (UID: \"fbfdd647-1d64-4d35-9af2-6dee52b4c860\") " Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.513136 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fbfdd647-1d64-4d35-9af2-6dee52b4c860-config" (OuterVolumeSpecName: "config") pod "fbfdd647-1d64-4d35-9af2-6dee52b4c860" (UID: "fbfdd647-1d64-4d35-9af2-6dee52b4c860"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.513850 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fbfdd647-1d64-4d35-9af2-6dee52b4c860-client-ca" (OuterVolumeSpecName: "client-ca") pod "fbfdd647-1d64-4d35-9af2-6dee52b4c860" (UID: "fbfdd647-1d64-4d35-9af2-6dee52b4c860"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.527507 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbfdd647-1d64-4d35-9af2-6dee52b4c860-kube-api-access-srhd7" (OuterVolumeSpecName: "kube-api-access-srhd7") pod "fbfdd647-1d64-4d35-9af2-6dee52b4c860" (UID: "fbfdd647-1d64-4d35-9af2-6dee52b4c860"). InnerVolumeSpecName "kube-api-access-srhd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.532298 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbfdd647-1d64-4d35-9af2-6dee52b4c860-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "fbfdd647-1d64-4d35-9af2-6dee52b4c860" (UID: "fbfdd647-1d64-4d35-9af2-6dee52b4c860"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.546582 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.613181 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c2d0611-58f8-4a7e-8280-361c80d62802-config\") pod \"3c2d0611-58f8-4a7e-8280-361c80d62802\" (UID: \"3c2d0611-58f8-4a7e-8280-361c80d62802\") " Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.613249 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3c2d0611-58f8-4a7e-8280-361c80d62802-proxy-ca-bundles\") pod \"3c2d0611-58f8-4a7e-8280-361c80d62802\" (UID: \"3c2d0611-58f8-4a7e-8280-361c80d62802\") " Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.613316 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pk8rh\" (UniqueName: \"kubernetes.io/projected/3c2d0611-58f8-4a7e-8280-361c80d62802-kube-api-access-pk8rh\") pod \"3c2d0611-58f8-4a7e-8280-361c80d62802\" (UID: \"3c2d0611-58f8-4a7e-8280-361c80d62802\") " Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.613340 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c2d0611-58f8-4a7e-8280-361c80d62802-serving-cert\") pod \"3c2d0611-58f8-4a7e-8280-361c80d62802\" (UID: \"3c2d0611-58f8-4a7e-8280-361c80d62802\") " Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.613493 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3c2d0611-58f8-4a7e-8280-361c80d62802-client-ca\") pod \"3c2d0611-58f8-4a7e-8280-361c80d62802\" (UID: \"3c2d0611-58f8-4a7e-8280-361c80d62802\") " Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.613846 4714 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fbfdd647-1d64-4d35-9af2-6dee52b4c860-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.613869 4714 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbfdd647-1d64-4d35-9af2-6dee52b4c860-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.613881 4714 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fbfdd647-1d64-4d35-9af2-6dee52b4c860-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.613895 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srhd7\" (UniqueName: \"kubernetes.io/projected/fbfdd647-1d64-4d35-9af2-6dee52b4c860-kube-api-access-srhd7\") on node \"crc\" DevicePath \"\"" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.614218 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c2d0611-58f8-4a7e-8280-361c80d62802-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "3c2d0611-58f8-4a7e-8280-361c80d62802" (UID: "3c2d0611-58f8-4a7e-8280-361c80d62802"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.614582 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c2d0611-58f8-4a7e-8280-361c80d62802-client-ca" (OuterVolumeSpecName: "client-ca") pod "3c2d0611-58f8-4a7e-8280-361c80d62802" (UID: "3c2d0611-58f8-4a7e-8280-361c80d62802"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.616121 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c2d0611-58f8-4a7e-8280-361c80d62802-config" (OuterVolumeSpecName: "config") pod "3c2d0611-58f8-4a7e-8280-361c80d62802" (UID: "3c2d0611-58f8-4a7e-8280-361c80d62802"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.618568 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c2d0611-58f8-4a7e-8280-361c80d62802-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3c2d0611-58f8-4a7e-8280-361c80d62802" (UID: "3c2d0611-58f8-4a7e-8280-361c80d62802"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.618712 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c2d0611-58f8-4a7e-8280-361c80d62802-kube-api-access-pk8rh" (OuterVolumeSpecName: "kube-api-access-pk8rh") pod "3c2d0611-58f8-4a7e-8280-361c80d62802" (UID: "3c2d0611-58f8-4a7e-8280-361c80d62802"). InnerVolumeSpecName "kube-api-access-pk8rh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.714912 4714 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3c2d0611-58f8-4a7e-8280-361c80d62802-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.714960 4714 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c2d0611-58f8-4a7e-8280-361c80d62802-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.714970 4714 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3c2d0611-58f8-4a7e-8280-361c80d62802-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.714981 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pk8rh\" (UniqueName: \"kubernetes.io/projected/3c2d0611-58f8-4a7e-8280-361c80d62802-kube-api-access-pk8rh\") on node \"crc\" DevicePath \"\"" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.714990 4714 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c2d0611-58f8-4a7e-8280-361c80d62802-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.868106 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29495055-4hxfc"] Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.983680 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29495055-4hxfc" event={"ID":"d0d972c0-6998-401f-8f0a-5bea6ed5590f","Type":"ContainerStarted","Data":"b6b3b048229369338cb18b65251b8d1da3ff36603f5a4de17d64b736eefdc448"} Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.985009 4714 generic.go:334] "Generic (PLEG): container finished" podID="3c2d0611-58f8-4a7e-8280-361c80d62802" containerID="1a48e7995b6652b74e320bf25807bcc51cd5f49496615ac9028d0bf40f37019e" exitCode=0 Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.985052 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" event={"ID":"3c2d0611-58f8-4a7e-8280-361c80d62802","Type":"ContainerDied","Data":"1a48e7995b6652b74e320bf25807bcc51cd5f49496615ac9028d0bf40f37019e"} Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.985093 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.985120 4714 scope.go:117] "RemoveContainer" containerID="1a48e7995b6652b74e320bf25807bcc51cd5f49496615ac9028d0bf40f37019e" Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.985101 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xlczd" event={"ID":"3c2d0611-58f8-4a7e-8280-361c80d62802","Type":"ContainerDied","Data":"997bb46f3e8548114daabdb0676e47c164f03b6651e1e3ef03b31f66106dbebd"} Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.988270 4714 generic.go:334] "Generic (PLEG): container finished" podID="fbfdd647-1d64-4d35-9af2-6dee52b4c860" containerID="3e427856d41457853f68d5a32dc2f87168c0f52da592bc0a2b3db34f710fea2b" exitCode=0 Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.988309 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" event={"ID":"fbfdd647-1d64-4d35-9af2-6dee52b4c860","Type":"ContainerDied","Data":"3e427856d41457853f68d5a32dc2f87168c0f52da592bc0a2b3db34f710fea2b"} Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.988332 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" event={"ID":"fbfdd647-1d64-4d35-9af2-6dee52b4c860","Type":"ContainerDied","Data":"79940598fef6f2445dc05d94ab28a7d984953a342201b3331c2b27e4796135a0"} Jan 29 16:15:28 crc kubenswrapper[4714]: I0129 16:15:28.988376 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw" Jan 29 16:15:29 crc kubenswrapper[4714]: I0129 16:15:29.004640 4714 scope.go:117] "RemoveContainer" containerID="1a48e7995b6652b74e320bf25807bcc51cd5f49496615ac9028d0bf40f37019e" Jan 29 16:15:29 crc kubenswrapper[4714]: E0129 16:15:29.005258 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a48e7995b6652b74e320bf25807bcc51cd5f49496615ac9028d0bf40f37019e\": container with ID starting with 1a48e7995b6652b74e320bf25807bcc51cd5f49496615ac9028d0bf40f37019e not found: ID does not exist" containerID="1a48e7995b6652b74e320bf25807bcc51cd5f49496615ac9028d0bf40f37019e" Jan 29 16:15:29 crc kubenswrapper[4714]: I0129 16:15:29.005304 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a48e7995b6652b74e320bf25807bcc51cd5f49496615ac9028d0bf40f37019e"} err="failed to get container status \"1a48e7995b6652b74e320bf25807bcc51cd5f49496615ac9028d0bf40f37019e\": rpc error: code = NotFound desc = could not find container \"1a48e7995b6652b74e320bf25807bcc51cd5f49496615ac9028d0bf40f37019e\": container with ID starting with 1a48e7995b6652b74e320bf25807bcc51cd5f49496615ac9028d0bf40f37019e not found: ID does not exist" Jan 29 16:15:29 crc kubenswrapper[4714]: I0129 16:15:29.005333 4714 scope.go:117] "RemoveContainer" containerID="3e427856d41457853f68d5a32dc2f87168c0f52da592bc0a2b3db34f710fea2b" Jan 29 16:15:29 crc kubenswrapper[4714]: I0129 16:15:29.025296 4714 scope.go:117] "RemoveContainer" containerID="3e427856d41457853f68d5a32dc2f87168c0f52da592bc0a2b3db34f710fea2b" Jan 29 16:15:29 crc kubenswrapper[4714]: E0129 16:15:29.028661 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e427856d41457853f68d5a32dc2f87168c0f52da592bc0a2b3db34f710fea2b\": container with ID starting with 3e427856d41457853f68d5a32dc2f87168c0f52da592bc0a2b3db34f710fea2b not found: ID does not exist" containerID="3e427856d41457853f68d5a32dc2f87168c0f52da592bc0a2b3db34f710fea2b" Jan 29 16:15:29 crc kubenswrapper[4714]: I0129 16:15:29.028719 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e427856d41457853f68d5a32dc2f87168c0f52da592bc0a2b3db34f710fea2b"} err="failed to get container status \"3e427856d41457853f68d5a32dc2f87168c0f52da592bc0a2b3db34f710fea2b\": rpc error: code = NotFound desc = could not find container \"3e427856d41457853f68d5a32dc2f87168c0f52da592bc0a2b3db34f710fea2b\": container with ID starting with 3e427856d41457853f68d5a32dc2f87168c0f52da592bc0a2b3db34f710fea2b not found: ID does not exist" Jan 29 16:15:29 crc kubenswrapper[4714]: I0129 16:15:29.034879 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw"] Jan 29 16:15:29 crc kubenswrapper[4714]: I0129 16:15:29.043272 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2qxw"] Jan 29 16:15:29 crc kubenswrapper[4714]: I0129 16:15:29.046584 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xlczd"] Jan 29 16:15:29 crc kubenswrapper[4714]: I0129 16:15:29.050207 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xlczd"] Jan 29 16:15:29 crc kubenswrapper[4714]: I0129 16:15:29.997180 4714 generic.go:334] "Generic (PLEG): container finished" podID="d0d972c0-6998-401f-8f0a-5bea6ed5590f" containerID="4b672d789aed1cc6acdb183556fd62ebd76f645741d01b1b70856eabac3f1f5f" exitCode=0 Jan 29 16:15:29 crc kubenswrapper[4714]: I0129 16:15:29.997253 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29495055-4hxfc" event={"ID":"d0d972c0-6998-401f-8f0a-5bea6ed5590f","Type":"ContainerDied","Data":"4b672d789aed1cc6acdb183556fd62ebd76f645741d01b1b70856eabac3f1f5f"} Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.009865 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6b89b4855f-vq5x7"] Jan 29 16:15:30 crc kubenswrapper[4714]: E0129 16:15:30.010221 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbfdd647-1d64-4d35-9af2-6dee52b4c860" containerName="route-controller-manager" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.010249 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbfdd647-1d64-4d35-9af2-6dee52b4c860" containerName="route-controller-manager" Jan 29 16:15:30 crc kubenswrapper[4714]: E0129 16:15:30.010278 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c2d0611-58f8-4a7e-8280-361c80d62802" containerName="controller-manager" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.010291 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c2d0611-58f8-4a7e-8280-361c80d62802" containerName="controller-manager" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.010511 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c2d0611-58f8-4a7e-8280-361c80d62802" containerName="controller-manager" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.010542 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbfdd647-1d64-4d35-9af2-6dee52b4c860" containerName="route-controller-manager" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.011127 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.014588 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.015650 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.016204 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.016365 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.016658 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.017178 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td"] Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.018357 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.019092 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.021722 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.022159 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.022823 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.023258 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.023312 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.025496 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.028845 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b89b4855f-vq5x7"] Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.031525 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.043954 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-866ps\" (UniqueName: \"kubernetes.io/projected/5304db45-6305-4176-a472-ce79c6a873bc-kube-api-access-866ps\") pod \"controller-manager-6b89b4855f-vq5x7\" (UID: \"5304db45-6305-4176-a472-ce79c6a873bc\") " pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.044006 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5304db45-6305-4176-a472-ce79c6a873bc-serving-cert\") pod \"controller-manager-6b89b4855f-vq5x7\" (UID: \"5304db45-6305-4176-a472-ce79c6a873bc\") " pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.044038 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5304db45-6305-4176-a472-ce79c6a873bc-proxy-ca-bundles\") pod \"controller-manager-6b89b4855f-vq5x7\" (UID: \"5304db45-6305-4176-a472-ce79c6a873bc\") " pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.046958 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5304db45-6305-4176-a472-ce79c6a873bc-client-ca\") pod \"controller-manager-6b89b4855f-vq5x7\" (UID: \"5304db45-6305-4176-a472-ce79c6a873bc\") " pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.047046 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5304db45-6305-4176-a472-ce79c6a873bc-config\") pod \"controller-manager-6b89b4855f-vq5x7\" (UID: \"5304db45-6305-4176-a472-ce79c6a873bc\") " pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.053960 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td"] Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.148565 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5304db45-6305-4176-a472-ce79c6a873bc-client-ca\") pod \"controller-manager-6b89b4855f-vq5x7\" (UID: \"5304db45-6305-4176-a472-ce79c6a873bc\") " pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.148625 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99c8ea80-a927-49e6-96fb-40c16f486883-serving-cert\") pod \"route-controller-manager-5ccd555bf-xt5td\" (UID: \"99c8ea80-a927-49e6-96fb-40c16f486883\") " pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.148653 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5304db45-6305-4176-a472-ce79c6a873bc-config\") pod \"controller-manager-6b89b4855f-vq5x7\" (UID: \"5304db45-6305-4176-a472-ce79c6a873bc\") " pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.148670 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/99c8ea80-a927-49e6-96fb-40c16f486883-client-ca\") pod \"route-controller-manager-5ccd555bf-xt5td\" (UID: \"99c8ea80-a927-49e6-96fb-40c16f486883\") " pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.148732 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk544\" (UniqueName: \"kubernetes.io/projected/99c8ea80-a927-49e6-96fb-40c16f486883-kube-api-access-gk544\") pod \"route-controller-manager-5ccd555bf-xt5td\" (UID: \"99c8ea80-a927-49e6-96fb-40c16f486883\") " pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.148762 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-866ps\" (UniqueName: \"kubernetes.io/projected/5304db45-6305-4176-a472-ce79c6a873bc-kube-api-access-866ps\") pod \"controller-manager-6b89b4855f-vq5x7\" (UID: \"5304db45-6305-4176-a472-ce79c6a873bc\") " pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.148778 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5304db45-6305-4176-a472-ce79c6a873bc-serving-cert\") pod \"controller-manager-6b89b4855f-vq5x7\" (UID: \"5304db45-6305-4176-a472-ce79c6a873bc\") " pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.148791 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5304db45-6305-4176-a472-ce79c6a873bc-proxy-ca-bundles\") pod \"controller-manager-6b89b4855f-vq5x7\" (UID: \"5304db45-6305-4176-a472-ce79c6a873bc\") " pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.148812 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c8ea80-a927-49e6-96fb-40c16f486883-config\") pod \"route-controller-manager-5ccd555bf-xt5td\" (UID: \"99c8ea80-a927-49e6-96fb-40c16f486883\") " pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.149837 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5304db45-6305-4176-a472-ce79c6a873bc-client-ca\") pod \"controller-manager-6b89b4855f-vq5x7\" (UID: \"5304db45-6305-4176-a472-ce79c6a873bc\") " pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.149971 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5304db45-6305-4176-a472-ce79c6a873bc-proxy-ca-bundles\") pod \"controller-manager-6b89b4855f-vq5x7\" (UID: \"5304db45-6305-4176-a472-ce79c6a873bc\") " pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.151130 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5304db45-6305-4176-a472-ce79c6a873bc-config\") pod \"controller-manager-6b89b4855f-vq5x7\" (UID: \"5304db45-6305-4176-a472-ce79c6a873bc\") " pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.160007 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5304db45-6305-4176-a472-ce79c6a873bc-serving-cert\") pod \"controller-manager-6b89b4855f-vq5x7\" (UID: \"5304db45-6305-4176-a472-ce79c6a873bc\") " pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.166430 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-866ps\" (UniqueName: \"kubernetes.io/projected/5304db45-6305-4176-a472-ce79c6a873bc-kube-api-access-866ps\") pod \"controller-manager-6b89b4855f-vq5x7\" (UID: \"5304db45-6305-4176-a472-ce79c6a873bc\") " pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.191578 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c2d0611-58f8-4a7e-8280-361c80d62802" path="/var/lib/kubelet/pods/3c2d0611-58f8-4a7e-8280-361c80d62802/volumes" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.192678 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbfdd647-1d64-4d35-9af2-6dee52b4c860" path="/var/lib/kubelet/pods/fbfdd647-1d64-4d35-9af2-6dee52b4c860/volumes" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.249874 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk544\" (UniqueName: \"kubernetes.io/projected/99c8ea80-a927-49e6-96fb-40c16f486883-kube-api-access-gk544\") pod \"route-controller-manager-5ccd555bf-xt5td\" (UID: \"99c8ea80-a927-49e6-96fb-40c16f486883\") " pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.249964 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c8ea80-a927-49e6-96fb-40c16f486883-config\") pod \"route-controller-manager-5ccd555bf-xt5td\" (UID: \"99c8ea80-a927-49e6-96fb-40c16f486883\") " pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.250020 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99c8ea80-a927-49e6-96fb-40c16f486883-serving-cert\") pod \"route-controller-manager-5ccd555bf-xt5td\" (UID: \"99c8ea80-a927-49e6-96fb-40c16f486883\") " pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.250051 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/99c8ea80-a927-49e6-96fb-40c16f486883-client-ca\") pod \"route-controller-manager-5ccd555bf-xt5td\" (UID: \"99c8ea80-a927-49e6-96fb-40c16f486883\") " pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.251039 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/99c8ea80-a927-49e6-96fb-40c16f486883-client-ca\") pod \"route-controller-manager-5ccd555bf-xt5td\" (UID: \"99c8ea80-a927-49e6-96fb-40c16f486883\") " pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.251271 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c8ea80-a927-49e6-96fb-40c16f486883-config\") pod \"route-controller-manager-5ccd555bf-xt5td\" (UID: \"99c8ea80-a927-49e6-96fb-40c16f486883\") " pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.253574 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99c8ea80-a927-49e6-96fb-40c16f486883-serving-cert\") pod \"route-controller-manager-5ccd555bf-xt5td\" (UID: \"99c8ea80-a927-49e6-96fb-40c16f486883\") " pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.264921 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk544\" (UniqueName: \"kubernetes.io/projected/99c8ea80-a927-49e6-96fb-40c16f486883-kube-api-access-gk544\") pod \"route-controller-manager-5ccd555bf-xt5td\" (UID: \"99c8ea80-a927-49e6-96fb-40c16f486883\") " pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.341703 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.364008 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.792254 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td"] Jan 29 16:15:30 crc kubenswrapper[4714]: I0129 16:15:30.803098 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b89b4855f-vq5x7"] Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.009736 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" event={"ID":"99c8ea80-a927-49e6-96fb-40c16f486883","Type":"ContainerStarted","Data":"d19fbbf2f1739d674126fe1890b29729a713b61f7c1394a35442defb39bc3f82"} Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.009783 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" event={"ID":"99c8ea80-a927-49e6-96fb-40c16f486883","Type":"ContainerStarted","Data":"af0e31a2974ff0d8ba2387e7344483b384496c48296610f8600e9690dddb022f"} Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.010022 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.014000 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" event={"ID":"5304db45-6305-4176-a472-ce79c6a873bc","Type":"ContainerStarted","Data":"68cf4b129bb68130c5704ef391f7af99e04fab51cea46cd69c6b56c7ab855d3d"} Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.014074 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" event={"ID":"5304db45-6305-4176-a472-ce79c6a873bc","Type":"ContainerStarted","Data":"51a48632aa0b57e7418c6d7638342a8c28a8562d8cca769a5ff82f52cb8a2256"} Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.014261 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.015179 4714 patch_prober.go:28] interesting pod/controller-manager-6b89b4855f-vq5x7 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" start-of-body= Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.015227 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" podUID="5304db45-6305-4176-a472-ce79c6a873bc" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.026766 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" podStartSLOduration=3.026747259 podStartE2EDuration="3.026747259s" podCreationTimestamp="2026-01-29 16:15:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:15:31.025198095 +0000 UTC m=+337.545699215" watchObservedRunningTime="2026-01-29 16:15:31.026747259 +0000 UTC m=+337.547248379" Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.052283 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" podStartSLOduration=3.052259729 podStartE2EDuration="3.052259729s" podCreationTimestamp="2026-01-29 16:15:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:15:31.046245429 +0000 UTC m=+337.566746549" watchObservedRunningTime="2026-01-29 16:15:31.052259729 +0000 UTC m=+337.572760849" Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.349197 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.365089 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495055-4hxfc" Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.468702 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0d972c0-6998-401f-8f0a-5bea6ed5590f-secret-volume\") pod \"d0d972c0-6998-401f-8f0a-5bea6ed5590f\" (UID: \"d0d972c0-6998-401f-8f0a-5bea6ed5590f\") " Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.468777 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjwmg\" (UniqueName: \"kubernetes.io/projected/d0d972c0-6998-401f-8f0a-5bea6ed5590f-kube-api-access-sjwmg\") pod \"d0d972c0-6998-401f-8f0a-5bea6ed5590f\" (UID: \"d0d972c0-6998-401f-8f0a-5bea6ed5590f\") " Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.468824 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0d972c0-6998-401f-8f0a-5bea6ed5590f-config-volume\") pod \"d0d972c0-6998-401f-8f0a-5bea6ed5590f\" (UID: \"d0d972c0-6998-401f-8f0a-5bea6ed5590f\") " Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.469439 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0d972c0-6998-401f-8f0a-5bea6ed5590f-config-volume" (OuterVolumeSpecName: "config-volume") pod "d0d972c0-6998-401f-8f0a-5bea6ed5590f" (UID: "d0d972c0-6998-401f-8f0a-5bea6ed5590f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.473714 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0d972c0-6998-401f-8f0a-5bea6ed5590f-kube-api-access-sjwmg" (OuterVolumeSpecName: "kube-api-access-sjwmg") pod "d0d972c0-6998-401f-8f0a-5bea6ed5590f" (UID: "d0d972c0-6998-401f-8f0a-5bea6ed5590f"). InnerVolumeSpecName "kube-api-access-sjwmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.477094 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0d972c0-6998-401f-8f0a-5bea6ed5590f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d0d972c0-6998-401f-8f0a-5bea6ed5590f" (UID: "d0d972c0-6998-401f-8f0a-5bea6ed5590f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.570380 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjwmg\" (UniqueName: \"kubernetes.io/projected/d0d972c0-6998-401f-8f0a-5bea6ed5590f-kube-api-access-sjwmg\") on node \"crc\" DevicePath \"\"" Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.570415 4714 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0d972c0-6998-401f-8f0a-5bea6ed5590f-config-volume\") on node \"crc\" DevicePath \"\"" Jan 29 16:15:31 crc kubenswrapper[4714]: I0129 16:15:31.570424 4714 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0d972c0-6998-401f-8f0a-5bea6ed5590f-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 29 16:15:32 crc kubenswrapper[4714]: I0129 16:15:32.023892 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29495055-4hxfc" event={"ID":"d0d972c0-6998-401f-8f0a-5bea6ed5590f","Type":"ContainerDied","Data":"b6b3b048229369338cb18b65251b8d1da3ff36603f5a4de17d64b736eefdc448"} Jan 29 16:15:32 crc kubenswrapper[4714]: I0129 16:15:32.024004 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6b3b048229369338cb18b65251b8d1da3ff36603f5a4de17d64b736eefdc448" Jan 29 16:15:32 crc kubenswrapper[4714]: I0129 16:15:32.024050 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495055-4hxfc" Jan 29 16:15:32 crc kubenswrapper[4714]: I0129 16:15:32.028373 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6b89b4855f-vq5x7" Jan 29 16:15:57 crc kubenswrapper[4714]: I0129 16:15:57.844761 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:15:57 crc kubenswrapper[4714]: I0129 16:15:57.845311 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:16:03 crc kubenswrapper[4714]: I0129 16:16:03.619896 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6bjgq"] Jan 29 16:16:03 crc kubenswrapper[4714]: I0129 16:16:03.622497 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6bjgq" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" containerName="registry-server" containerID="cri-o://7d09f4f35a658993558ef15a141e74b57af657ad6538fc1eecadfe107e507595" gracePeriod=2 Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.058716 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6bjgq" Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.166083 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98a35d03-ef3b-4341-9866-56d12a28aee3-catalog-content\") pod \"98a35d03-ef3b-4341-9866-56d12a28aee3\" (UID: \"98a35d03-ef3b-4341-9866-56d12a28aee3\") " Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.166169 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98a35d03-ef3b-4341-9866-56d12a28aee3-utilities\") pod \"98a35d03-ef3b-4341-9866-56d12a28aee3\" (UID: \"98a35d03-ef3b-4341-9866-56d12a28aee3\") " Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.166314 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gks4z\" (UniqueName: \"kubernetes.io/projected/98a35d03-ef3b-4341-9866-56d12a28aee3-kube-api-access-gks4z\") pod \"98a35d03-ef3b-4341-9866-56d12a28aee3\" (UID: \"98a35d03-ef3b-4341-9866-56d12a28aee3\") " Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.167321 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98a35d03-ef3b-4341-9866-56d12a28aee3-utilities" (OuterVolumeSpecName: "utilities") pod "98a35d03-ef3b-4341-9866-56d12a28aee3" (UID: "98a35d03-ef3b-4341-9866-56d12a28aee3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.173369 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98a35d03-ef3b-4341-9866-56d12a28aee3-kube-api-access-gks4z" (OuterVolumeSpecName: "kube-api-access-gks4z") pod "98a35d03-ef3b-4341-9866-56d12a28aee3" (UID: "98a35d03-ef3b-4341-9866-56d12a28aee3"). InnerVolumeSpecName "kube-api-access-gks4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.221534 4714 generic.go:334] "Generic (PLEG): container finished" podID="98a35d03-ef3b-4341-9866-56d12a28aee3" containerID="7d09f4f35a658993558ef15a141e74b57af657ad6538fc1eecadfe107e507595" exitCode=0 Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.221581 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6bjgq" Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.221603 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bjgq" event={"ID":"98a35d03-ef3b-4341-9866-56d12a28aee3","Type":"ContainerDied","Data":"7d09f4f35a658993558ef15a141e74b57af657ad6538fc1eecadfe107e507595"} Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.221781 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bjgq" event={"ID":"98a35d03-ef3b-4341-9866-56d12a28aee3","Type":"ContainerDied","Data":"d38f58d434dcb4497833c894d88b3ceb4be10c7a4d69f2a5403bda7aa069a88c"} Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.221831 4714 scope.go:117] "RemoveContainer" containerID="7d09f4f35a658993558ef15a141e74b57af657ad6538fc1eecadfe107e507595" Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.236190 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98a35d03-ef3b-4341-9866-56d12a28aee3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98a35d03-ef3b-4341-9866-56d12a28aee3" (UID: "98a35d03-ef3b-4341-9866-56d12a28aee3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.245042 4714 scope.go:117] "RemoveContainer" containerID="aaa814f95896ac1eaedc7b86e2197c8140345f5296d9cb5ecb6a3dac9cee5ab6" Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.268307 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gks4z\" (UniqueName: \"kubernetes.io/projected/98a35d03-ef3b-4341-9866-56d12a28aee3-kube-api-access-gks4z\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.268343 4714 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98a35d03-ef3b-4341-9866-56d12a28aee3-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.268355 4714 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98a35d03-ef3b-4341-9866-56d12a28aee3-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.271405 4714 scope.go:117] "RemoveContainer" containerID="7d5da21ddc846f7484a829e4e1b7d4f27ddad6196e5ccbce162fd0a2651869dd" Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.290816 4714 scope.go:117] "RemoveContainer" containerID="7d09f4f35a658993558ef15a141e74b57af657ad6538fc1eecadfe107e507595" Jan 29 16:16:04 crc kubenswrapper[4714]: E0129 16:16:04.291371 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d09f4f35a658993558ef15a141e74b57af657ad6538fc1eecadfe107e507595\": container with ID starting with 7d09f4f35a658993558ef15a141e74b57af657ad6538fc1eecadfe107e507595 not found: ID does not exist" containerID="7d09f4f35a658993558ef15a141e74b57af657ad6538fc1eecadfe107e507595" Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.291429 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d09f4f35a658993558ef15a141e74b57af657ad6538fc1eecadfe107e507595"} err="failed to get container status \"7d09f4f35a658993558ef15a141e74b57af657ad6538fc1eecadfe107e507595\": rpc error: code = NotFound desc = could not find container \"7d09f4f35a658993558ef15a141e74b57af657ad6538fc1eecadfe107e507595\": container with ID starting with 7d09f4f35a658993558ef15a141e74b57af657ad6538fc1eecadfe107e507595 not found: ID does not exist" Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.291482 4714 scope.go:117] "RemoveContainer" containerID="aaa814f95896ac1eaedc7b86e2197c8140345f5296d9cb5ecb6a3dac9cee5ab6" Jan 29 16:16:04 crc kubenswrapper[4714]: E0129 16:16:04.291897 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aaa814f95896ac1eaedc7b86e2197c8140345f5296d9cb5ecb6a3dac9cee5ab6\": container with ID starting with aaa814f95896ac1eaedc7b86e2197c8140345f5296d9cb5ecb6a3dac9cee5ab6 not found: ID does not exist" containerID="aaa814f95896ac1eaedc7b86e2197c8140345f5296d9cb5ecb6a3dac9cee5ab6" Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.291991 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaa814f95896ac1eaedc7b86e2197c8140345f5296d9cb5ecb6a3dac9cee5ab6"} err="failed to get container status \"aaa814f95896ac1eaedc7b86e2197c8140345f5296d9cb5ecb6a3dac9cee5ab6\": rpc error: code = NotFound desc = could not find container \"aaa814f95896ac1eaedc7b86e2197c8140345f5296d9cb5ecb6a3dac9cee5ab6\": container with ID starting with aaa814f95896ac1eaedc7b86e2197c8140345f5296d9cb5ecb6a3dac9cee5ab6 not found: ID does not exist" Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.292030 4714 scope.go:117] "RemoveContainer" containerID="7d5da21ddc846f7484a829e4e1b7d4f27ddad6196e5ccbce162fd0a2651869dd" Jan 29 16:16:04 crc kubenswrapper[4714]: E0129 16:16:04.292353 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d5da21ddc846f7484a829e4e1b7d4f27ddad6196e5ccbce162fd0a2651869dd\": container with ID starting with 7d5da21ddc846f7484a829e4e1b7d4f27ddad6196e5ccbce162fd0a2651869dd not found: ID does not exist" containerID="7d5da21ddc846f7484a829e4e1b7d4f27ddad6196e5ccbce162fd0a2651869dd" Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.292386 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d5da21ddc846f7484a829e4e1b7d4f27ddad6196e5ccbce162fd0a2651869dd"} err="failed to get container status \"7d5da21ddc846f7484a829e4e1b7d4f27ddad6196e5ccbce162fd0a2651869dd\": rpc error: code = NotFound desc = could not find container \"7d5da21ddc846f7484a829e4e1b7d4f27ddad6196e5ccbce162fd0a2651869dd\": container with ID starting with 7d5da21ddc846f7484a829e4e1b7d4f27ddad6196e5ccbce162fd0a2651869dd not found: ID does not exist" Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.553337 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6bjgq"] Jan 29 16:16:04 crc kubenswrapper[4714]: I0129 16:16:04.557165 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6bjgq"] Jan 29 16:16:06 crc kubenswrapper[4714]: I0129 16:16:06.197026 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" path="/var/lib/kubelet/pods/98a35d03-ef3b-4341-9866-56d12a28aee3/volumes" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.726024 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xrw9s"] Jan 29 16:16:09 crc kubenswrapper[4714]: E0129 16:16:09.726509 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" containerName="registry-server" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.726524 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" containerName="registry-server" Jan 29 16:16:09 crc kubenswrapper[4714]: E0129 16:16:09.726537 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" containerName="extract-content" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.726545 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" containerName="extract-content" Jan 29 16:16:09 crc kubenswrapper[4714]: E0129 16:16:09.726563 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0d972c0-6998-401f-8f0a-5bea6ed5590f" containerName="collect-profiles" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.726572 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0d972c0-6998-401f-8f0a-5bea6ed5590f" containerName="collect-profiles" Jan 29 16:16:09 crc kubenswrapper[4714]: E0129 16:16:09.726582 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" containerName="extract-utilities" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.726590 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" containerName="extract-utilities" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.726729 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0d972c0-6998-401f-8f0a-5bea6ed5590f" containerName="collect-profiles" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.726748 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="98a35d03-ef3b-4341-9866-56d12a28aee3" containerName="registry-server" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.727175 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.741470 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xrw9s"] Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.841573 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5f131437-8551-46b5-b9e4-ca30784e8e76-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.841649 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5f131437-8551-46b5-b9e4-ca30784e8e76-registry-certificates\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.841880 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5f131437-8551-46b5-b9e4-ca30784e8e76-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.841952 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5f131437-8551-46b5-b9e4-ca30784e8e76-registry-tls\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.842048 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5f131437-8551-46b5-b9e4-ca30784e8e76-trusted-ca\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.842095 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5f131437-8551-46b5-b9e4-ca30784e8e76-bound-sa-token\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.842128 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csr5w\" (UniqueName: \"kubernetes.io/projected/5f131437-8551-46b5-b9e4-ca30784e8e76-kube-api-access-csr5w\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.842183 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.869565 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.943693 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5f131437-8551-46b5-b9e4-ca30784e8e76-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.943769 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5f131437-8551-46b5-b9e4-ca30784e8e76-registry-certificates\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.943854 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5f131437-8551-46b5-b9e4-ca30784e8e76-registry-tls\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.943890 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5f131437-8551-46b5-b9e4-ca30784e8e76-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.943983 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5f131437-8551-46b5-b9e4-ca30784e8e76-trusted-ca\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.944033 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5f131437-8551-46b5-b9e4-ca30784e8e76-bound-sa-token\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.944568 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csr5w\" (UniqueName: \"kubernetes.io/projected/5f131437-8551-46b5-b9e4-ca30784e8e76-kube-api-access-csr5w\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.945870 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5f131437-8551-46b5-b9e4-ca30784e8e76-registry-certificates\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.946331 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5f131437-8551-46b5-b9e4-ca30784e8e76-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.946518 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5f131437-8551-46b5-b9e4-ca30784e8e76-trusted-ca\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.952093 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5f131437-8551-46b5-b9e4-ca30784e8e76-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.954013 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5f131437-8551-46b5-b9e4-ca30784e8e76-registry-tls\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.965363 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csr5w\" (UniqueName: \"kubernetes.io/projected/5f131437-8551-46b5-b9e4-ca30784e8e76-kube-api-access-csr5w\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:09 crc kubenswrapper[4714]: I0129 16:16:09.965642 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5f131437-8551-46b5-b9e4-ca30784e8e76-bound-sa-token\") pod \"image-registry-66df7c8f76-xrw9s\" (UID: \"5f131437-8551-46b5-b9e4-ca30784e8e76\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:10 crc kubenswrapper[4714]: I0129 16:16:10.047268 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:10 crc kubenswrapper[4714]: I0129 16:16:10.258443 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td"] Jan 29 16:16:10 crc kubenswrapper[4714]: I0129 16:16:10.263669 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" podUID="99c8ea80-a927-49e6-96fb-40c16f486883" containerName="route-controller-manager" containerID="cri-o://d19fbbf2f1739d674126fe1890b29729a713b61f7c1394a35442defb39bc3f82" gracePeriod=30 Jan 29 16:16:10 crc kubenswrapper[4714]: I0129 16:16:10.347794 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xrw9s"] Jan 29 16:16:10 crc kubenswrapper[4714]: I0129 16:16:10.365306 4714 patch_prober.go:28] interesting pod/route-controller-manager-5ccd555bf-xt5td container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.60:8443/healthz\": dial tcp 10.217.0.60:8443: connect: connection refused" start-of-body= Jan 29 16:16:10 crc kubenswrapper[4714]: I0129 16:16:10.365356 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" podUID="99c8ea80-a927-49e6-96fb-40c16f486883" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.60:8443/healthz\": dial tcp 10.217.0.60:8443: connect: connection refused" Jan 29 16:16:10 crc kubenswrapper[4714]: I0129 16:16:10.601802 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" Jan 29 16:16:10 crc kubenswrapper[4714]: I0129 16:16:10.756799 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/99c8ea80-a927-49e6-96fb-40c16f486883-client-ca\") pod \"99c8ea80-a927-49e6-96fb-40c16f486883\" (UID: \"99c8ea80-a927-49e6-96fb-40c16f486883\") " Jan 29 16:16:10 crc kubenswrapper[4714]: I0129 16:16:10.756873 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c8ea80-a927-49e6-96fb-40c16f486883-config\") pod \"99c8ea80-a927-49e6-96fb-40c16f486883\" (UID: \"99c8ea80-a927-49e6-96fb-40c16f486883\") " Jan 29 16:16:10 crc kubenswrapper[4714]: I0129 16:16:10.757018 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gk544\" (UniqueName: \"kubernetes.io/projected/99c8ea80-a927-49e6-96fb-40c16f486883-kube-api-access-gk544\") pod \"99c8ea80-a927-49e6-96fb-40c16f486883\" (UID: \"99c8ea80-a927-49e6-96fb-40c16f486883\") " Jan 29 16:16:10 crc kubenswrapper[4714]: I0129 16:16:10.757178 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99c8ea80-a927-49e6-96fb-40c16f486883-serving-cert\") pod \"99c8ea80-a927-49e6-96fb-40c16f486883\" (UID: \"99c8ea80-a927-49e6-96fb-40c16f486883\") " Jan 29 16:16:10 crc kubenswrapper[4714]: I0129 16:16:10.757623 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99c8ea80-a927-49e6-96fb-40c16f486883-config" (OuterVolumeSpecName: "config") pod "99c8ea80-a927-49e6-96fb-40c16f486883" (UID: "99c8ea80-a927-49e6-96fb-40c16f486883"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:16:10 crc kubenswrapper[4714]: I0129 16:16:10.757621 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99c8ea80-a927-49e6-96fb-40c16f486883-client-ca" (OuterVolumeSpecName: "client-ca") pod "99c8ea80-a927-49e6-96fb-40c16f486883" (UID: "99c8ea80-a927-49e6-96fb-40c16f486883"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:16:10 crc kubenswrapper[4714]: I0129 16:16:10.758541 4714 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/99c8ea80-a927-49e6-96fb-40c16f486883-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:10 crc kubenswrapper[4714]: I0129 16:16:10.758580 4714 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c8ea80-a927-49e6-96fb-40c16f486883-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:10 crc kubenswrapper[4714]: I0129 16:16:10.761466 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99c8ea80-a927-49e6-96fb-40c16f486883-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "99c8ea80-a927-49e6-96fb-40c16f486883" (UID: "99c8ea80-a927-49e6-96fb-40c16f486883"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:16:10 crc kubenswrapper[4714]: I0129 16:16:10.761515 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99c8ea80-a927-49e6-96fb-40c16f486883-kube-api-access-gk544" (OuterVolumeSpecName: "kube-api-access-gk544") pod "99c8ea80-a927-49e6-96fb-40c16f486883" (UID: "99c8ea80-a927-49e6-96fb-40c16f486883"). InnerVolumeSpecName "kube-api-access-gk544". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:16:10 crc kubenswrapper[4714]: I0129 16:16:10.859495 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gk544\" (UniqueName: \"kubernetes.io/projected/99c8ea80-a927-49e6-96fb-40c16f486883-kube-api-access-gk544\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:10 crc kubenswrapper[4714]: I0129 16:16:10.859529 4714 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99c8ea80-a927-49e6-96fb-40c16f486883-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:11 crc kubenswrapper[4714]: I0129 16:16:11.283810 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" event={"ID":"5f131437-8551-46b5-b9e4-ca30784e8e76","Type":"ContainerStarted","Data":"51f4e55276037a86c65e6532c8fb92ea992ee6083c96ae7e2613baadc4ab112e"} Jan 29 16:16:11 crc kubenswrapper[4714]: I0129 16:16:11.284281 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:11 crc kubenswrapper[4714]: I0129 16:16:11.284344 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" event={"ID":"5f131437-8551-46b5-b9e4-ca30784e8e76","Type":"ContainerStarted","Data":"208ee0164257d5ce848beee5714cbcaa9c52741aa6c8b5e3e845659b1c10dd3c"} Jan 29 16:16:11 crc kubenswrapper[4714]: I0129 16:16:11.285390 4714 generic.go:334] "Generic (PLEG): container finished" podID="99c8ea80-a927-49e6-96fb-40c16f486883" containerID="d19fbbf2f1739d674126fe1890b29729a713b61f7c1394a35442defb39bc3f82" exitCode=0 Jan 29 16:16:11 crc kubenswrapper[4714]: I0129 16:16:11.285442 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" event={"ID":"99c8ea80-a927-49e6-96fb-40c16f486883","Type":"ContainerDied","Data":"d19fbbf2f1739d674126fe1890b29729a713b61f7c1394a35442defb39bc3f82"} Jan 29 16:16:11 crc kubenswrapper[4714]: I0129 16:16:11.285475 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" Jan 29 16:16:11 crc kubenswrapper[4714]: I0129 16:16:11.285504 4714 scope.go:117] "RemoveContainer" containerID="d19fbbf2f1739d674126fe1890b29729a713b61f7c1394a35442defb39bc3f82" Jan 29 16:16:11 crc kubenswrapper[4714]: I0129 16:16:11.285486 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td" event={"ID":"99c8ea80-a927-49e6-96fb-40c16f486883","Type":"ContainerDied","Data":"af0e31a2974ff0d8ba2387e7344483b384496c48296610f8600e9690dddb022f"} Jan 29 16:16:11 crc kubenswrapper[4714]: I0129 16:16:11.304703 4714 scope.go:117] "RemoveContainer" containerID="d19fbbf2f1739d674126fe1890b29729a713b61f7c1394a35442defb39bc3f82" Jan 29 16:16:11 crc kubenswrapper[4714]: E0129 16:16:11.305210 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d19fbbf2f1739d674126fe1890b29729a713b61f7c1394a35442defb39bc3f82\": container with ID starting with d19fbbf2f1739d674126fe1890b29729a713b61f7c1394a35442defb39bc3f82 not found: ID does not exist" containerID="d19fbbf2f1739d674126fe1890b29729a713b61f7c1394a35442defb39bc3f82" Jan 29 16:16:11 crc kubenswrapper[4714]: I0129 16:16:11.305295 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d19fbbf2f1739d674126fe1890b29729a713b61f7c1394a35442defb39bc3f82"} err="failed to get container status \"d19fbbf2f1739d674126fe1890b29729a713b61f7c1394a35442defb39bc3f82\": rpc error: code = NotFound desc = could not find container \"d19fbbf2f1739d674126fe1890b29729a713b61f7c1394a35442defb39bc3f82\": container with ID starting with d19fbbf2f1739d674126fe1890b29729a713b61f7c1394a35442defb39bc3f82 not found: ID does not exist" Jan 29 16:16:11 crc kubenswrapper[4714]: I0129 16:16:11.312982 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" podStartSLOduration=2.31296472 podStartE2EDuration="2.31296472s" podCreationTimestamp="2026-01-29 16:16:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:16:11.308521404 +0000 UTC m=+377.829022524" watchObservedRunningTime="2026-01-29 16:16:11.31296472 +0000 UTC m=+377.833465840" Jan 29 16:16:11 crc kubenswrapper[4714]: I0129 16:16:11.334602 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td"] Jan 29 16:16:11 crc kubenswrapper[4714]: I0129 16:16:11.341264 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5ccd555bf-xt5td"] Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.051373 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl"] Jan 29 16:16:12 crc kubenswrapper[4714]: E0129 16:16:12.051635 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99c8ea80-a927-49e6-96fb-40c16f486883" containerName="route-controller-manager" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.051651 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="99c8ea80-a927-49e6-96fb-40c16f486883" containerName="route-controller-manager" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.051769 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="99c8ea80-a927-49e6-96fb-40c16f486883" containerName="route-controller-manager" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.052277 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.054339 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.054870 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.054876 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.054976 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.055290 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.056292 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.065264 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl"] Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.179399 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw9mv\" (UniqueName: \"kubernetes.io/projected/cd27896d-ddee-4374-ad3b-dca8f6f7f5d4-kube-api-access-jw9mv\") pod \"route-controller-manager-6c6f6c447c-4rzrl\" (UID: \"cd27896d-ddee-4374-ad3b-dca8f6f7f5d4\") " pod="openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.179484 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cd27896d-ddee-4374-ad3b-dca8f6f7f5d4-client-ca\") pod \"route-controller-manager-6c6f6c447c-4rzrl\" (UID: \"cd27896d-ddee-4374-ad3b-dca8f6f7f5d4\") " pod="openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.179570 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd27896d-ddee-4374-ad3b-dca8f6f7f5d4-config\") pod \"route-controller-manager-6c6f6c447c-4rzrl\" (UID: \"cd27896d-ddee-4374-ad3b-dca8f6f7f5d4\") " pod="openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.179877 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd27896d-ddee-4374-ad3b-dca8f6f7f5d4-serving-cert\") pod \"route-controller-manager-6c6f6c447c-4rzrl\" (UID: \"cd27896d-ddee-4374-ad3b-dca8f6f7f5d4\") " pod="openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.192667 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99c8ea80-a927-49e6-96fb-40c16f486883" path="/var/lib/kubelet/pods/99c8ea80-a927-49e6-96fb-40c16f486883/volumes" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.281845 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd27896d-ddee-4374-ad3b-dca8f6f7f5d4-config\") pod \"route-controller-manager-6c6f6c447c-4rzrl\" (UID: \"cd27896d-ddee-4374-ad3b-dca8f6f7f5d4\") " pod="openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.282026 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd27896d-ddee-4374-ad3b-dca8f6f7f5d4-serving-cert\") pod \"route-controller-manager-6c6f6c447c-4rzrl\" (UID: \"cd27896d-ddee-4374-ad3b-dca8f6f7f5d4\") " pod="openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.282078 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw9mv\" (UniqueName: \"kubernetes.io/projected/cd27896d-ddee-4374-ad3b-dca8f6f7f5d4-kube-api-access-jw9mv\") pod \"route-controller-manager-6c6f6c447c-4rzrl\" (UID: \"cd27896d-ddee-4374-ad3b-dca8f6f7f5d4\") " pod="openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.282139 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cd27896d-ddee-4374-ad3b-dca8f6f7f5d4-client-ca\") pod \"route-controller-manager-6c6f6c447c-4rzrl\" (UID: \"cd27896d-ddee-4374-ad3b-dca8f6f7f5d4\") " pod="openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.284171 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cd27896d-ddee-4374-ad3b-dca8f6f7f5d4-client-ca\") pod \"route-controller-manager-6c6f6c447c-4rzrl\" (UID: \"cd27896d-ddee-4374-ad3b-dca8f6f7f5d4\") " pod="openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.285297 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd27896d-ddee-4374-ad3b-dca8f6f7f5d4-config\") pod \"route-controller-manager-6c6f6c447c-4rzrl\" (UID: \"cd27896d-ddee-4374-ad3b-dca8f6f7f5d4\") " pod="openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.290768 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd27896d-ddee-4374-ad3b-dca8f6f7f5d4-serving-cert\") pod \"route-controller-manager-6c6f6c447c-4rzrl\" (UID: \"cd27896d-ddee-4374-ad3b-dca8f6f7f5d4\") " pod="openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.319631 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw9mv\" (UniqueName: \"kubernetes.io/projected/cd27896d-ddee-4374-ad3b-dca8f6f7f5d4-kube-api-access-jw9mv\") pod \"route-controller-manager-6c6f6c447c-4rzrl\" (UID: \"cd27896d-ddee-4374-ad3b-dca8f6f7f5d4\") " pod="openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.379843 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl" Jan 29 16:16:12 crc kubenswrapper[4714]: I0129 16:16:12.679898 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl"] Jan 29 16:16:12 crc kubenswrapper[4714]: W0129 16:16:12.690230 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd27896d_ddee_4374_ad3b_dca8f6f7f5d4.slice/crio-4be1e858e531185043040e9e09cb296c9b0d2208caa7180ed7beaebb666fa3ef WatchSource:0}: Error finding container 4be1e858e531185043040e9e09cb296c9b0d2208caa7180ed7beaebb666fa3ef: Status 404 returned error can't find the container with id 4be1e858e531185043040e9e09cb296c9b0d2208caa7180ed7beaebb666fa3ef Jan 29 16:16:13 crc kubenswrapper[4714]: I0129 16:16:13.302798 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl" event={"ID":"cd27896d-ddee-4374-ad3b-dca8f6f7f5d4","Type":"ContainerStarted","Data":"6e0400d6dd5564863c84909aa18ec4f8123768e7a6efb155e6816fa23ca4a8de"} Jan 29 16:16:13 crc kubenswrapper[4714]: I0129 16:16:13.303250 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl" event={"ID":"cd27896d-ddee-4374-ad3b-dca8f6f7f5d4","Type":"ContainerStarted","Data":"4be1e858e531185043040e9e09cb296c9b0d2208caa7180ed7beaebb666fa3ef"} Jan 29 16:16:13 crc kubenswrapper[4714]: I0129 16:16:13.335876 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl" podStartSLOduration=3.335838791 podStartE2EDuration="3.335838791s" podCreationTimestamp="2026-01-29 16:16:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:16:13.329511332 +0000 UTC m=+379.850012452" watchObservedRunningTime="2026-01-29 16:16:13.335838791 +0000 UTC m=+379.856339951" Jan 29 16:16:14 crc kubenswrapper[4714]: I0129 16:16:14.309006 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl" Jan 29 16:16:14 crc kubenswrapper[4714]: I0129 16:16:14.315027 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6c6f6c447c-4rzrl" Jan 29 16:16:23 crc kubenswrapper[4714]: I0129 16:16:23.806159 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-74twj"] Jan 29 16:16:23 crc kubenswrapper[4714]: I0129 16:16:23.807139 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-74twj" podUID="a97ed1ff-657f-4bde-943b-78caf9d07f92" containerName="registry-server" containerID="cri-o://0725e5fe8581c9f5ab88fa8ad4af11d5f996f4972602031ec88175b5fda32f79" gracePeriod=30 Jan 29 16:16:23 crc kubenswrapper[4714]: I0129 16:16:23.817809 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xtr82"] Jan 29 16:16:23 crc kubenswrapper[4714]: I0129 16:16:23.818142 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xtr82" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" containerName="registry-server" containerID="cri-o://a2c143c1c72b06e6085dcb21f057a3e45d817ac3a8bf8d7e3516db54610f130e" gracePeriod=30 Jan 29 16:16:23 crc kubenswrapper[4714]: I0129 16:16:23.826389 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-l2t56"] Jan 29 16:16:23 crc kubenswrapper[4714]: I0129 16:16:23.826851 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" podUID="80515d06-c09e-4c9d-a90f-43cc84edf4c9" containerName="marketplace-operator" containerID="cri-o://2cfb1164c8a5f24d11bd2a23214b6a7408be50990447c790085e11ea6faaec50" gracePeriod=30 Jan 29 16:16:23 crc kubenswrapper[4714]: I0129 16:16:23.840345 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7rvrl"] Jan 29 16:16:23 crc kubenswrapper[4714]: I0129 16:16:23.841304 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7rvrl" Jan 29 16:16:23 crc kubenswrapper[4714]: I0129 16:16:23.844790 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nssrv"] Jan 29 16:16:23 crc kubenswrapper[4714]: I0129 16:16:23.845107 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nssrv" podUID="eae853ba-61c9-439b-9dc9-21567075f18a" containerName="registry-server" containerID="cri-o://52559e4420e032272c3033326a96872ff005d794cf95b2dce2bffa130b6cf392" gracePeriod=30 Jan 29 16:16:23 crc kubenswrapper[4714]: I0129 16:16:23.865858 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lb68h"] Jan 29 16:16:23 crc kubenswrapper[4714]: I0129 16:16:23.866485 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lb68h" podUID="d05e7c79-7d66-4453-aedb-f240784ff294" containerName="registry-server" containerID="cri-o://25adf524f9c5473bfa242fd63827380ee76b9012a8eb32666464c503ff1ae5f5" gracePeriod=30 Jan 29 16:16:23 crc kubenswrapper[4714]: I0129 16:16:23.874213 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7rvrl"] Jan 29 16:16:23 crc kubenswrapper[4714]: I0129 16:16:23.970976 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2696757f-83ca-42df-9855-f76adeee02bb-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7rvrl\" (UID: \"2696757f-83ca-42df-9855-f76adeee02bb\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rvrl" Jan 29 16:16:23 crc kubenswrapper[4714]: I0129 16:16:23.971174 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4dlv\" (UniqueName: \"kubernetes.io/projected/2696757f-83ca-42df-9855-f76adeee02bb-kube-api-access-m4dlv\") pod \"marketplace-operator-79b997595-7rvrl\" (UID: \"2696757f-83ca-42df-9855-f76adeee02bb\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rvrl" Jan 29 16:16:23 crc kubenswrapper[4714]: I0129 16:16:23.971267 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2696757f-83ca-42df-9855-f76adeee02bb-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7rvrl\" (UID: \"2696757f-83ca-42df-9855-f76adeee02bb\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rvrl" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.073884 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2696757f-83ca-42df-9855-f76adeee02bb-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7rvrl\" (UID: \"2696757f-83ca-42df-9855-f76adeee02bb\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rvrl" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.073993 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2696757f-83ca-42df-9855-f76adeee02bb-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7rvrl\" (UID: \"2696757f-83ca-42df-9855-f76adeee02bb\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rvrl" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.074058 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4dlv\" (UniqueName: \"kubernetes.io/projected/2696757f-83ca-42df-9855-f76adeee02bb-kube-api-access-m4dlv\") pod \"marketplace-operator-79b997595-7rvrl\" (UID: \"2696757f-83ca-42df-9855-f76adeee02bb\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rvrl" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.075603 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2696757f-83ca-42df-9855-f76adeee02bb-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7rvrl\" (UID: \"2696757f-83ca-42df-9855-f76adeee02bb\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rvrl" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.089361 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2696757f-83ca-42df-9855-f76adeee02bb-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7rvrl\" (UID: \"2696757f-83ca-42df-9855-f76adeee02bb\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rvrl" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.107300 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4dlv\" (UniqueName: \"kubernetes.io/projected/2696757f-83ca-42df-9855-f76adeee02bb-kube-api-access-m4dlv\") pod \"marketplace-operator-79b997595-7rvrl\" (UID: \"2696757f-83ca-42df-9855-f76adeee02bb\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rvrl" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.242298 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7rvrl" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.258475 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xtr82" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.262816 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-74twj" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.268633 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.352774 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nssrv" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.361438 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lb68h" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.383616 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11a30de8-b234-47b4-8fd0-44f0c428be78-utilities\") pod \"11a30de8-b234-47b4-8fd0-44f0c428be78\" (UID: \"11a30de8-b234-47b4-8fd0-44f0c428be78\") " Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.383699 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a97ed1ff-657f-4bde-943b-78caf9d07f92-catalog-content\") pod \"a97ed1ff-657f-4bde-943b-78caf9d07f92\" (UID: \"a97ed1ff-657f-4bde-943b-78caf9d07f92\") " Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.383793 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11a30de8-b234-47b4-8fd0-44f0c428be78-catalog-content\") pod \"11a30de8-b234-47b4-8fd0-44f0c428be78\" (UID: \"11a30de8-b234-47b4-8fd0-44f0c428be78\") " Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.383822 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a97ed1ff-657f-4bde-943b-78caf9d07f92-utilities\") pod \"a97ed1ff-657f-4bde-943b-78caf9d07f92\" (UID: \"a97ed1ff-657f-4bde-943b-78caf9d07f92\") " Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.383874 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbn22\" (UniqueName: \"kubernetes.io/projected/11a30de8-b234-47b4-8fd0-44f0c428be78-kube-api-access-zbn22\") pod \"11a30de8-b234-47b4-8fd0-44f0c428be78\" (UID: \"11a30de8-b234-47b4-8fd0-44f0c428be78\") " Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.383904 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/80515d06-c09e-4c9d-a90f-43cc84edf4c9-marketplace-operator-metrics\") pod \"80515d06-c09e-4c9d-a90f-43cc84edf4c9\" (UID: \"80515d06-c09e-4c9d-a90f-43cc84edf4c9\") " Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.383960 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8bfv\" (UniqueName: \"kubernetes.io/projected/a97ed1ff-657f-4bde-943b-78caf9d07f92-kube-api-access-v8bfv\") pod \"a97ed1ff-657f-4bde-943b-78caf9d07f92\" (UID: \"a97ed1ff-657f-4bde-943b-78caf9d07f92\") " Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.383990 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfc66\" (UniqueName: \"kubernetes.io/projected/80515d06-c09e-4c9d-a90f-43cc84edf4c9-kube-api-access-xfc66\") pod \"80515d06-c09e-4c9d-a90f-43cc84edf4c9\" (UID: \"80515d06-c09e-4c9d-a90f-43cc84edf4c9\") " Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.384067 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80515d06-c09e-4c9d-a90f-43cc84edf4c9-marketplace-trusted-ca\") pod \"80515d06-c09e-4c9d-a90f-43cc84edf4c9\" (UID: \"80515d06-c09e-4c9d-a90f-43cc84edf4c9\") " Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.384871 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11a30de8-b234-47b4-8fd0-44f0c428be78-utilities" (OuterVolumeSpecName: "utilities") pod "11a30de8-b234-47b4-8fd0-44f0c428be78" (UID: "11a30de8-b234-47b4-8fd0-44f0c428be78"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.385770 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80515d06-c09e-4c9d-a90f-43cc84edf4c9-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "80515d06-c09e-4c9d-a90f-43cc84edf4c9" (UID: "80515d06-c09e-4c9d-a90f-43cc84edf4c9"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.385965 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a97ed1ff-657f-4bde-943b-78caf9d07f92-utilities" (OuterVolumeSpecName: "utilities") pod "a97ed1ff-657f-4bde-943b-78caf9d07f92" (UID: "a97ed1ff-657f-4bde-943b-78caf9d07f92"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.389618 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11a30de8-b234-47b4-8fd0-44f0c428be78-kube-api-access-zbn22" (OuterVolumeSpecName: "kube-api-access-zbn22") pod "11a30de8-b234-47b4-8fd0-44f0c428be78" (UID: "11a30de8-b234-47b4-8fd0-44f0c428be78"). InnerVolumeSpecName "kube-api-access-zbn22". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.390865 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a97ed1ff-657f-4bde-943b-78caf9d07f92-kube-api-access-v8bfv" (OuterVolumeSpecName: "kube-api-access-v8bfv") pod "a97ed1ff-657f-4bde-943b-78caf9d07f92" (UID: "a97ed1ff-657f-4bde-943b-78caf9d07f92"). InnerVolumeSpecName "kube-api-access-v8bfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.407329 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80515d06-c09e-4c9d-a90f-43cc84edf4c9-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "80515d06-c09e-4c9d-a90f-43cc84edf4c9" (UID: "80515d06-c09e-4c9d-a90f-43cc84edf4c9"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.410003 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80515d06-c09e-4c9d-a90f-43cc84edf4c9-kube-api-access-xfc66" (OuterVolumeSpecName: "kube-api-access-xfc66") pod "80515d06-c09e-4c9d-a90f-43cc84edf4c9" (UID: "80515d06-c09e-4c9d-a90f-43cc84edf4c9"). InnerVolumeSpecName "kube-api-access-xfc66". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.411702 4714 generic.go:334] "Generic (PLEG): container finished" podID="11a30de8-b234-47b4-8fd0-44f0c428be78" containerID="a2c143c1c72b06e6085dcb21f057a3e45d817ac3a8bf8d7e3516db54610f130e" exitCode=0 Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.411792 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xtr82" event={"ID":"11a30de8-b234-47b4-8fd0-44f0c428be78","Type":"ContainerDied","Data":"a2c143c1c72b06e6085dcb21f057a3e45d817ac3a8bf8d7e3516db54610f130e"} Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.411823 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xtr82" event={"ID":"11a30de8-b234-47b4-8fd0-44f0c428be78","Type":"ContainerDied","Data":"d1e11cf94d1ae7d280d25746da20bca8871b9f9c8323efe87d1cfb324504d7a1"} Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.411842 4714 scope.go:117] "RemoveContainer" containerID="a2c143c1c72b06e6085dcb21f057a3e45d817ac3a8bf8d7e3516db54610f130e" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.412121 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xtr82" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.419319 4714 generic.go:334] "Generic (PLEG): container finished" podID="a97ed1ff-657f-4bde-943b-78caf9d07f92" containerID="0725e5fe8581c9f5ab88fa8ad4af11d5f996f4972602031ec88175b5fda32f79" exitCode=0 Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.419457 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-74twj" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.419599 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-74twj" event={"ID":"a97ed1ff-657f-4bde-943b-78caf9d07f92","Type":"ContainerDied","Data":"0725e5fe8581c9f5ab88fa8ad4af11d5f996f4972602031ec88175b5fda32f79"} Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.419672 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-74twj" event={"ID":"a97ed1ff-657f-4bde-943b-78caf9d07f92","Type":"ContainerDied","Data":"27e514e7925336355503e562c2b866089bbb8f20f6235853c55635bfeebcfe8c"} Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.426202 4714 generic.go:334] "Generic (PLEG): container finished" podID="d05e7c79-7d66-4453-aedb-f240784ff294" containerID="25adf524f9c5473bfa242fd63827380ee76b9012a8eb32666464c503ff1ae5f5" exitCode=0 Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.426293 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lb68h" event={"ID":"d05e7c79-7d66-4453-aedb-f240784ff294","Type":"ContainerDied","Data":"25adf524f9c5473bfa242fd63827380ee76b9012a8eb32666464c503ff1ae5f5"} Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.426347 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lb68h" event={"ID":"d05e7c79-7d66-4453-aedb-f240784ff294","Type":"ContainerDied","Data":"5e22f2e727671a2879c86dcb9146aebbe76ddedf77fd5e705c834b21cf8bd941"} Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.426449 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lb68h" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.429181 4714 generic.go:334] "Generic (PLEG): container finished" podID="80515d06-c09e-4c9d-a90f-43cc84edf4c9" containerID="2cfb1164c8a5f24d11bd2a23214b6a7408be50990447c790085e11ea6faaec50" exitCode=0 Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.429232 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.429256 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" event={"ID":"80515d06-c09e-4c9d-a90f-43cc84edf4c9","Type":"ContainerDied","Data":"2cfb1164c8a5f24d11bd2a23214b6a7408be50990447c790085e11ea6faaec50"} Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.429295 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-l2t56" event={"ID":"80515d06-c09e-4c9d-a90f-43cc84edf4c9","Type":"ContainerDied","Data":"5880f1855bae3fd6f603655d40b770623f038db9a3cb9db3918877f801567acc"} Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.431495 4714 generic.go:334] "Generic (PLEG): container finished" podID="eae853ba-61c9-439b-9dc9-21567075f18a" containerID="52559e4420e032272c3033326a96872ff005d794cf95b2dce2bffa130b6cf392" exitCode=0 Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.431530 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nssrv" event={"ID":"eae853ba-61c9-439b-9dc9-21567075f18a","Type":"ContainerDied","Data":"52559e4420e032272c3033326a96872ff005d794cf95b2dce2bffa130b6cf392"} Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.431555 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nssrv" event={"ID":"eae853ba-61c9-439b-9dc9-21567075f18a","Type":"ContainerDied","Data":"11eca2d99e975c8d4c6d498c418a6ed86174580092ad733d4cf31d057f9d974e"} Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.431616 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nssrv" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.452783 4714 scope.go:117] "RemoveContainer" containerID="8a54a45940dea4f60292ad1cc53a3fa404d30031632cdd928c54b5c498a0d947" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.461050 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-l2t56"] Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.464726 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-l2t56"] Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.466222 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a97ed1ff-657f-4bde-943b-78caf9d07f92-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a97ed1ff-657f-4bde-943b-78caf9d07f92" (UID: "a97ed1ff-657f-4bde-943b-78caf9d07f92"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.469818 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11a30de8-b234-47b4-8fd0-44f0c428be78-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "11a30de8-b234-47b4-8fd0-44f0c428be78" (UID: "11a30de8-b234-47b4-8fd0-44f0c428be78"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.472824 4714 scope.go:117] "RemoveContainer" containerID="70b956b273a4676b7f7a5e461f43c09f1067f834a7d754d57cae60e15152821d" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.485521 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntklh\" (UniqueName: \"kubernetes.io/projected/eae853ba-61c9-439b-9dc9-21567075f18a-kube-api-access-ntklh\") pod \"eae853ba-61c9-439b-9dc9-21567075f18a\" (UID: \"eae853ba-61c9-439b-9dc9-21567075f18a\") " Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.485587 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d05e7c79-7d66-4453-aedb-f240784ff294-utilities\") pod \"d05e7c79-7d66-4453-aedb-f240784ff294\" (UID: \"d05e7c79-7d66-4453-aedb-f240784ff294\") " Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.485615 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d05e7c79-7d66-4453-aedb-f240784ff294-catalog-content\") pod \"d05e7c79-7d66-4453-aedb-f240784ff294\" (UID: \"d05e7c79-7d66-4453-aedb-f240784ff294\") " Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.485658 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae853ba-61c9-439b-9dc9-21567075f18a-utilities\") pod \"eae853ba-61c9-439b-9dc9-21567075f18a\" (UID: \"eae853ba-61c9-439b-9dc9-21567075f18a\") " Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.485695 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae853ba-61c9-439b-9dc9-21567075f18a-catalog-content\") pod \"eae853ba-61c9-439b-9dc9-21567075f18a\" (UID: \"eae853ba-61c9-439b-9dc9-21567075f18a\") " Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.485737 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4jjj\" (UniqueName: \"kubernetes.io/projected/d05e7c79-7d66-4453-aedb-f240784ff294-kube-api-access-m4jjj\") pod \"d05e7c79-7d66-4453-aedb-f240784ff294\" (UID: \"d05e7c79-7d66-4453-aedb-f240784ff294\") " Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.485980 4714 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11a30de8-b234-47b4-8fd0-44f0c428be78-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.485997 4714 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a97ed1ff-657f-4bde-943b-78caf9d07f92-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.486010 4714 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11a30de8-b234-47b4-8fd0-44f0c428be78-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.486022 4714 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a97ed1ff-657f-4bde-943b-78caf9d07f92-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.486032 4714 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/80515d06-c09e-4c9d-a90f-43cc84edf4c9-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.486041 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbn22\" (UniqueName: \"kubernetes.io/projected/11a30de8-b234-47b4-8fd0-44f0c428be78-kube-api-access-zbn22\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.486049 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8bfv\" (UniqueName: \"kubernetes.io/projected/a97ed1ff-657f-4bde-943b-78caf9d07f92-kube-api-access-v8bfv\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.486058 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfc66\" (UniqueName: \"kubernetes.io/projected/80515d06-c09e-4c9d-a90f-43cc84edf4c9-kube-api-access-xfc66\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.486069 4714 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80515d06-c09e-4c9d-a90f-43cc84edf4c9-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.488395 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d05e7c79-7d66-4453-aedb-f240784ff294-utilities" (OuterVolumeSpecName: "utilities") pod "d05e7c79-7d66-4453-aedb-f240784ff294" (UID: "d05e7c79-7d66-4453-aedb-f240784ff294"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.488544 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eae853ba-61c9-439b-9dc9-21567075f18a-utilities" (OuterVolumeSpecName: "utilities") pod "eae853ba-61c9-439b-9dc9-21567075f18a" (UID: "eae853ba-61c9-439b-9dc9-21567075f18a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.491335 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eae853ba-61c9-439b-9dc9-21567075f18a-kube-api-access-ntklh" (OuterVolumeSpecName: "kube-api-access-ntklh") pod "eae853ba-61c9-439b-9dc9-21567075f18a" (UID: "eae853ba-61c9-439b-9dc9-21567075f18a"). InnerVolumeSpecName "kube-api-access-ntklh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.492030 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d05e7c79-7d66-4453-aedb-f240784ff294-kube-api-access-m4jjj" (OuterVolumeSpecName: "kube-api-access-m4jjj") pod "d05e7c79-7d66-4453-aedb-f240784ff294" (UID: "d05e7c79-7d66-4453-aedb-f240784ff294"). InnerVolumeSpecName "kube-api-access-m4jjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.508404 4714 scope.go:117] "RemoveContainer" containerID="a2c143c1c72b06e6085dcb21f057a3e45d817ac3a8bf8d7e3516db54610f130e" Jan 29 16:16:24 crc kubenswrapper[4714]: E0129 16:16:24.509022 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2c143c1c72b06e6085dcb21f057a3e45d817ac3a8bf8d7e3516db54610f130e\": container with ID starting with a2c143c1c72b06e6085dcb21f057a3e45d817ac3a8bf8d7e3516db54610f130e not found: ID does not exist" containerID="a2c143c1c72b06e6085dcb21f057a3e45d817ac3a8bf8d7e3516db54610f130e" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.509060 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2c143c1c72b06e6085dcb21f057a3e45d817ac3a8bf8d7e3516db54610f130e"} err="failed to get container status \"a2c143c1c72b06e6085dcb21f057a3e45d817ac3a8bf8d7e3516db54610f130e\": rpc error: code = NotFound desc = could not find container \"a2c143c1c72b06e6085dcb21f057a3e45d817ac3a8bf8d7e3516db54610f130e\": container with ID starting with a2c143c1c72b06e6085dcb21f057a3e45d817ac3a8bf8d7e3516db54610f130e not found: ID does not exist" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.509092 4714 scope.go:117] "RemoveContainer" containerID="8a54a45940dea4f60292ad1cc53a3fa404d30031632cdd928c54b5c498a0d947" Jan 29 16:16:24 crc kubenswrapper[4714]: E0129 16:16:24.509458 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a54a45940dea4f60292ad1cc53a3fa404d30031632cdd928c54b5c498a0d947\": container with ID starting with 8a54a45940dea4f60292ad1cc53a3fa404d30031632cdd928c54b5c498a0d947 not found: ID does not exist" containerID="8a54a45940dea4f60292ad1cc53a3fa404d30031632cdd928c54b5c498a0d947" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.509473 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a54a45940dea4f60292ad1cc53a3fa404d30031632cdd928c54b5c498a0d947"} err="failed to get container status \"8a54a45940dea4f60292ad1cc53a3fa404d30031632cdd928c54b5c498a0d947\": rpc error: code = NotFound desc = could not find container \"8a54a45940dea4f60292ad1cc53a3fa404d30031632cdd928c54b5c498a0d947\": container with ID starting with 8a54a45940dea4f60292ad1cc53a3fa404d30031632cdd928c54b5c498a0d947 not found: ID does not exist" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.509486 4714 scope.go:117] "RemoveContainer" containerID="70b956b273a4676b7f7a5e461f43c09f1067f834a7d754d57cae60e15152821d" Jan 29 16:16:24 crc kubenswrapper[4714]: E0129 16:16:24.509805 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70b956b273a4676b7f7a5e461f43c09f1067f834a7d754d57cae60e15152821d\": container with ID starting with 70b956b273a4676b7f7a5e461f43c09f1067f834a7d754d57cae60e15152821d not found: ID does not exist" containerID="70b956b273a4676b7f7a5e461f43c09f1067f834a7d754d57cae60e15152821d" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.509837 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70b956b273a4676b7f7a5e461f43c09f1067f834a7d754d57cae60e15152821d"} err="failed to get container status \"70b956b273a4676b7f7a5e461f43c09f1067f834a7d754d57cae60e15152821d\": rpc error: code = NotFound desc = could not find container \"70b956b273a4676b7f7a5e461f43c09f1067f834a7d754d57cae60e15152821d\": container with ID starting with 70b956b273a4676b7f7a5e461f43c09f1067f834a7d754d57cae60e15152821d not found: ID does not exist" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.509853 4714 scope.go:117] "RemoveContainer" containerID="0725e5fe8581c9f5ab88fa8ad4af11d5f996f4972602031ec88175b5fda32f79" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.511783 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eae853ba-61c9-439b-9dc9-21567075f18a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eae853ba-61c9-439b-9dc9-21567075f18a" (UID: "eae853ba-61c9-439b-9dc9-21567075f18a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.525099 4714 scope.go:117] "RemoveContainer" containerID="b55a4a53101476467f28cefb11eb7b554ba1847628876fc42f62a75c8730a4f6" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.542971 4714 scope.go:117] "RemoveContainer" containerID="bebaa193fd909649d996ac5ecb12f75e7aa251dd1a4b1b911882734c87e4b046" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.557637 4714 scope.go:117] "RemoveContainer" containerID="0725e5fe8581c9f5ab88fa8ad4af11d5f996f4972602031ec88175b5fda32f79" Jan 29 16:16:24 crc kubenswrapper[4714]: E0129 16:16:24.558026 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0725e5fe8581c9f5ab88fa8ad4af11d5f996f4972602031ec88175b5fda32f79\": container with ID starting with 0725e5fe8581c9f5ab88fa8ad4af11d5f996f4972602031ec88175b5fda32f79 not found: ID does not exist" containerID="0725e5fe8581c9f5ab88fa8ad4af11d5f996f4972602031ec88175b5fda32f79" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.558068 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0725e5fe8581c9f5ab88fa8ad4af11d5f996f4972602031ec88175b5fda32f79"} err="failed to get container status \"0725e5fe8581c9f5ab88fa8ad4af11d5f996f4972602031ec88175b5fda32f79\": rpc error: code = NotFound desc = could not find container \"0725e5fe8581c9f5ab88fa8ad4af11d5f996f4972602031ec88175b5fda32f79\": container with ID starting with 0725e5fe8581c9f5ab88fa8ad4af11d5f996f4972602031ec88175b5fda32f79 not found: ID does not exist" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.558100 4714 scope.go:117] "RemoveContainer" containerID="b55a4a53101476467f28cefb11eb7b554ba1847628876fc42f62a75c8730a4f6" Jan 29 16:16:24 crc kubenswrapper[4714]: E0129 16:16:24.558380 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b55a4a53101476467f28cefb11eb7b554ba1847628876fc42f62a75c8730a4f6\": container with ID starting with b55a4a53101476467f28cefb11eb7b554ba1847628876fc42f62a75c8730a4f6 not found: ID does not exist" containerID="b55a4a53101476467f28cefb11eb7b554ba1847628876fc42f62a75c8730a4f6" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.558416 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b55a4a53101476467f28cefb11eb7b554ba1847628876fc42f62a75c8730a4f6"} err="failed to get container status \"b55a4a53101476467f28cefb11eb7b554ba1847628876fc42f62a75c8730a4f6\": rpc error: code = NotFound desc = could not find container \"b55a4a53101476467f28cefb11eb7b554ba1847628876fc42f62a75c8730a4f6\": container with ID starting with b55a4a53101476467f28cefb11eb7b554ba1847628876fc42f62a75c8730a4f6 not found: ID does not exist" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.558441 4714 scope.go:117] "RemoveContainer" containerID="bebaa193fd909649d996ac5ecb12f75e7aa251dd1a4b1b911882734c87e4b046" Jan 29 16:16:24 crc kubenswrapper[4714]: E0129 16:16:24.558721 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bebaa193fd909649d996ac5ecb12f75e7aa251dd1a4b1b911882734c87e4b046\": container with ID starting with bebaa193fd909649d996ac5ecb12f75e7aa251dd1a4b1b911882734c87e4b046 not found: ID does not exist" containerID="bebaa193fd909649d996ac5ecb12f75e7aa251dd1a4b1b911882734c87e4b046" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.558745 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bebaa193fd909649d996ac5ecb12f75e7aa251dd1a4b1b911882734c87e4b046"} err="failed to get container status \"bebaa193fd909649d996ac5ecb12f75e7aa251dd1a4b1b911882734c87e4b046\": rpc error: code = NotFound desc = could not find container \"bebaa193fd909649d996ac5ecb12f75e7aa251dd1a4b1b911882734c87e4b046\": container with ID starting with bebaa193fd909649d996ac5ecb12f75e7aa251dd1a4b1b911882734c87e4b046 not found: ID does not exist" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.558768 4714 scope.go:117] "RemoveContainer" containerID="25adf524f9c5473bfa242fd63827380ee76b9012a8eb32666464c503ff1ae5f5" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.574809 4714 scope.go:117] "RemoveContainer" containerID="dc4ea141b6a80ae098de7ca5a17e9e2b1ec3ebf2112f640c6fad4d5fcc75a51c" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.587381 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntklh\" (UniqueName: \"kubernetes.io/projected/eae853ba-61c9-439b-9dc9-21567075f18a-kube-api-access-ntklh\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.587408 4714 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d05e7c79-7d66-4453-aedb-f240784ff294-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.587419 4714 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae853ba-61c9-439b-9dc9-21567075f18a-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.587428 4714 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae853ba-61c9-439b-9dc9-21567075f18a-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.587436 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4jjj\" (UniqueName: \"kubernetes.io/projected/d05e7c79-7d66-4453-aedb-f240784ff294-kube-api-access-m4jjj\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.588770 4714 scope.go:117] "RemoveContainer" containerID="0e066cf92f43693eba9898b6ef36d8d3eb21b4fa7c877b99db7e3e39dddda134" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.602639 4714 scope.go:117] "RemoveContainer" containerID="25adf524f9c5473bfa242fd63827380ee76b9012a8eb32666464c503ff1ae5f5" Jan 29 16:16:24 crc kubenswrapper[4714]: E0129 16:16:24.604103 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25adf524f9c5473bfa242fd63827380ee76b9012a8eb32666464c503ff1ae5f5\": container with ID starting with 25adf524f9c5473bfa242fd63827380ee76b9012a8eb32666464c503ff1ae5f5 not found: ID does not exist" containerID="25adf524f9c5473bfa242fd63827380ee76b9012a8eb32666464c503ff1ae5f5" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.604141 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25adf524f9c5473bfa242fd63827380ee76b9012a8eb32666464c503ff1ae5f5"} err="failed to get container status \"25adf524f9c5473bfa242fd63827380ee76b9012a8eb32666464c503ff1ae5f5\": rpc error: code = NotFound desc = could not find container \"25adf524f9c5473bfa242fd63827380ee76b9012a8eb32666464c503ff1ae5f5\": container with ID starting with 25adf524f9c5473bfa242fd63827380ee76b9012a8eb32666464c503ff1ae5f5 not found: ID does not exist" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.604174 4714 scope.go:117] "RemoveContainer" containerID="dc4ea141b6a80ae098de7ca5a17e9e2b1ec3ebf2112f640c6fad4d5fcc75a51c" Jan 29 16:16:24 crc kubenswrapper[4714]: E0129 16:16:24.604523 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc4ea141b6a80ae098de7ca5a17e9e2b1ec3ebf2112f640c6fad4d5fcc75a51c\": container with ID starting with dc4ea141b6a80ae098de7ca5a17e9e2b1ec3ebf2112f640c6fad4d5fcc75a51c not found: ID does not exist" containerID="dc4ea141b6a80ae098de7ca5a17e9e2b1ec3ebf2112f640c6fad4d5fcc75a51c" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.604570 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc4ea141b6a80ae098de7ca5a17e9e2b1ec3ebf2112f640c6fad4d5fcc75a51c"} err="failed to get container status \"dc4ea141b6a80ae098de7ca5a17e9e2b1ec3ebf2112f640c6fad4d5fcc75a51c\": rpc error: code = NotFound desc = could not find container \"dc4ea141b6a80ae098de7ca5a17e9e2b1ec3ebf2112f640c6fad4d5fcc75a51c\": container with ID starting with dc4ea141b6a80ae098de7ca5a17e9e2b1ec3ebf2112f640c6fad4d5fcc75a51c not found: ID does not exist" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.604603 4714 scope.go:117] "RemoveContainer" containerID="0e066cf92f43693eba9898b6ef36d8d3eb21b4fa7c877b99db7e3e39dddda134" Jan 29 16:16:24 crc kubenswrapper[4714]: E0129 16:16:24.605309 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e066cf92f43693eba9898b6ef36d8d3eb21b4fa7c877b99db7e3e39dddda134\": container with ID starting with 0e066cf92f43693eba9898b6ef36d8d3eb21b4fa7c877b99db7e3e39dddda134 not found: ID does not exist" containerID="0e066cf92f43693eba9898b6ef36d8d3eb21b4fa7c877b99db7e3e39dddda134" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.605520 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e066cf92f43693eba9898b6ef36d8d3eb21b4fa7c877b99db7e3e39dddda134"} err="failed to get container status \"0e066cf92f43693eba9898b6ef36d8d3eb21b4fa7c877b99db7e3e39dddda134\": rpc error: code = NotFound desc = could not find container \"0e066cf92f43693eba9898b6ef36d8d3eb21b4fa7c877b99db7e3e39dddda134\": container with ID starting with 0e066cf92f43693eba9898b6ef36d8d3eb21b4fa7c877b99db7e3e39dddda134 not found: ID does not exist" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.605542 4714 scope.go:117] "RemoveContainer" containerID="2cfb1164c8a5f24d11bd2a23214b6a7408be50990447c790085e11ea6faaec50" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.614464 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d05e7c79-7d66-4453-aedb-f240784ff294-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d05e7c79-7d66-4453-aedb-f240784ff294" (UID: "d05e7c79-7d66-4453-aedb-f240784ff294"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.622881 4714 scope.go:117] "RemoveContainer" containerID="2cfb1164c8a5f24d11bd2a23214b6a7408be50990447c790085e11ea6faaec50" Jan 29 16:16:24 crc kubenswrapper[4714]: E0129 16:16:24.623691 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cfb1164c8a5f24d11bd2a23214b6a7408be50990447c790085e11ea6faaec50\": container with ID starting with 2cfb1164c8a5f24d11bd2a23214b6a7408be50990447c790085e11ea6faaec50 not found: ID does not exist" containerID="2cfb1164c8a5f24d11bd2a23214b6a7408be50990447c790085e11ea6faaec50" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.623755 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cfb1164c8a5f24d11bd2a23214b6a7408be50990447c790085e11ea6faaec50"} err="failed to get container status \"2cfb1164c8a5f24d11bd2a23214b6a7408be50990447c790085e11ea6faaec50\": rpc error: code = NotFound desc = could not find container \"2cfb1164c8a5f24d11bd2a23214b6a7408be50990447c790085e11ea6faaec50\": container with ID starting with 2cfb1164c8a5f24d11bd2a23214b6a7408be50990447c790085e11ea6faaec50 not found: ID does not exist" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.623800 4714 scope.go:117] "RemoveContainer" containerID="52559e4420e032272c3033326a96872ff005d794cf95b2dce2bffa130b6cf392" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.639088 4714 scope.go:117] "RemoveContainer" containerID="af26cf3e86e4001d9e7f83e8aa3b6ea940d03f589aa1d1907c488eb1df46568b" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.653700 4714 scope.go:117] "RemoveContainer" containerID="4ddfcc2a03d228c12293d91791425de7e03e1ce7bf286f993aef049f853f58fe" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.668871 4714 scope.go:117] "RemoveContainer" containerID="52559e4420e032272c3033326a96872ff005d794cf95b2dce2bffa130b6cf392" Jan 29 16:16:24 crc kubenswrapper[4714]: E0129 16:16:24.669665 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52559e4420e032272c3033326a96872ff005d794cf95b2dce2bffa130b6cf392\": container with ID starting with 52559e4420e032272c3033326a96872ff005d794cf95b2dce2bffa130b6cf392 not found: ID does not exist" containerID="52559e4420e032272c3033326a96872ff005d794cf95b2dce2bffa130b6cf392" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.669709 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52559e4420e032272c3033326a96872ff005d794cf95b2dce2bffa130b6cf392"} err="failed to get container status \"52559e4420e032272c3033326a96872ff005d794cf95b2dce2bffa130b6cf392\": rpc error: code = NotFound desc = could not find container \"52559e4420e032272c3033326a96872ff005d794cf95b2dce2bffa130b6cf392\": container with ID starting with 52559e4420e032272c3033326a96872ff005d794cf95b2dce2bffa130b6cf392 not found: ID does not exist" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.669743 4714 scope.go:117] "RemoveContainer" containerID="af26cf3e86e4001d9e7f83e8aa3b6ea940d03f589aa1d1907c488eb1df46568b" Jan 29 16:16:24 crc kubenswrapper[4714]: E0129 16:16:24.670081 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af26cf3e86e4001d9e7f83e8aa3b6ea940d03f589aa1d1907c488eb1df46568b\": container with ID starting with af26cf3e86e4001d9e7f83e8aa3b6ea940d03f589aa1d1907c488eb1df46568b not found: ID does not exist" containerID="af26cf3e86e4001d9e7f83e8aa3b6ea940d03f589aa1d1907c488eb1df46568b" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.670151 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af26cf3e86e4001d9e7f83e8aa3b6ea940d03f589aa1d1907c488eb1df46568b"} err="failed to get container status \"af26cf3e86e4001d9e7f83e8aa3b6ea940d03f589aa1d1907c488eb1df46568b\": rpc error: code = NotFound desc = could not find container \"af26cf3e86e4001d9e7f83e8aa3b6ea940d03f589aa1d1907c488eb1df46568b\": container with ID starting with af26cf3e86e4001d9e7f83e8aa3b6ea940d03f589aa1d1907c488eb1df46568b not found: ID does not exist" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.670197 4714 scope.go:117] "RemoveContainer" containerID="4ddfcc2a03d228c12293d91791425de7e03e1ce7bf286f993aef049f853f58fe" Jan 29 16:16:24 crc kubenswrapper[4714]: E0129 16:16:24.670527 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ddfcc2a03d228c12293d91791425de7e03e1ce7bf286f993aef049f853f58fe\": container with ID starting with 4ddfcc2a03d228c12293d91791425de7e03e1ce7bf286f993aef049f853f58fe not found: ID does not exist" containerID="4ddfcc2a03d228c12293d91791425de7e03e1ce7bf286f993aef049f853f58fe" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.670561 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ddfcc2a03d228c12293d91791425de7e03e1ce7bf286f993aef049f853f58fe"} err="failed to get container status \"4ddfcc2a03d228c12293d91791425de7e03e1ce7bf286f993aef049f853f58fe\": rpc error: code = NotFound desc = could not find container \"4ddfcc2a03d228c12293d91791425de7e03e1ce7bf286f993aef049f853f58fe\": container with ID starting with 4ddfcc2a03d228c12293d91791425de7e03e1ce7bf286f993aef049f853f58fe not found: ID does not exist" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.688631 4714 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d05e7c79-7d66-4453-aedb-f240784ff294-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.709372 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7rvrl"] Jan 29 16:16:24 crc kubenswrapper[4714]: W0129 16:16:24.712315 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2696757f_83ca_42df_9855_f76adeee02bb.slice/crio-cdd61689243a69bcd2c8fda2e4c87e8c55a3f1474ee1d4589fa09b3921bf8f48 WatchSource:0}: Error finding container cdd61689243a69bcd2c8fda2e4c87e8c55a3f1474ee1d4589fa09b3921bf8f48: Status 404 returned error can't find the container with id cdd61689243a69bcd2c8fda2e4c87e8c55a3f1474ee1d4589fa09b3921bf8f48 Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.755131 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xtr82"] Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.765726 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xtr82"] Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.787098 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lb68h"] Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.790871 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lb68h"] Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.815584 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nssrv"] Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.827458 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nssrv"] Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.831790 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-74twj"] Jan 29 16:16:24 crc kubenswrapper[4714]: I0129 16:16:24.835330 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-74twj"] Jan 29 16:16:25 crc kubenswrapper[4714]: I0129 16:16:25.454397 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7rvrl" event={"ID":"2696757f-83ca-42df-9855-f76adeee02bb","Type":"ContainerStarted","Data":"b1c9a9e34a8f6e1c92c31b01db09a13a690eb1c214066679034d3934a7c755a7"} Jan 29 16:16:25 crc kubenswrapper[4714]: I0129 16:16:25.454818 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7rvrl" event={"ID":"2696757f-83ca-42df-9855-f76adeee02bb","Type":"ContainerStarted","Data":"cdd61689243a69bcd2c8fda2e4c87e8c55a3f1474ee1d4589fa09b3921bf8f48"} Jan 29 16:16:25 crc kubenswrapper[4714]: I0129 16:16:25.455888 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-7rvrl" Jan 29 16:16:25 crc kubenswrapper[4714]: I0129 16:16:25.458984 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-7rvrl" Jan 29 16:16:25 crc kubenswrapper[4714]: I0129 16:16:25.478029 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-7rvrl" podStartSLOduration=2.478013781 podStartE2EDuration="2.478013781s" podCreationTimestamp="2026-01-29 16:16:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:16:25.477917528 +0000 UTC m=+391.998418688" watchObservedRunningTime="2026-01-29 16:16:25.478013781 +0000 UTC m=+391.998514901" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.195218 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" path="/var/lib/kubelet/pods/11a30de8-b234-47b4-8fd0-44f0c428be78/volumes" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.196852 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80515d06-c09e-4c9d-a90f-43cc84edf4c9" path="/var/lib/kubelet/pods/80515d06-c09e-4c9d-a90f-43cc84edf4c9/volumes" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.197764 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a97ed1ff-657f-4bde-943b-78caf9d07f92" path="/var/lib/kubelet/pods/a97ed1ff-657f-4bde-943b-78caf9d07f92/volumes" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.200569 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d05e7c79-7d66-4453-aedb-f240784ff294" path="/var/lib/kubelet/pods/d05e7c79-7d66-4453-aedb-f240784ff294/volumes" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.201991 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eae853ba-61c9-439b-9dc9-21567075f18a" path="/var/lib/kubelet/pods/eae853ba-61c9-439b-9dc9-21567075f18a/volumes" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.565257 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6gkpz"] Jan 29 16:16:26 crc kubenswrapper[4714]: E0129 16:16:26.565453 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" containerName="extract-utilities" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.565465 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" containerName="extract-utilities" Jan 29 16:16:26 crc kubenswrapper[4714]: E0129 16:16:26.565498 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a97ed1ff-657f-4bde-943b-78caf9d07f92" containerName="extract-content" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.565505 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="a97ed1ff-657f-4bde-943b-78caf9d07f92" containerName="extract-content" Jan 29 16:16:26 crc kubenswrapper[4714]: E0129 16:16:26.565517 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80515d06-c09e-4c9d-a90f-43cc84edf4c9" containerName="marketplace-operator" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.565523 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="80515d06-c09e-4c9d-a90f-43cc84edf4c9" containerName="marketplace-operator" Jan 29 16:16:26 crc kubenswrapper[4714]: E0129 16:16:26.565530 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a97ed1ff-657f-4bde-943b-78caf9d07f92" containerName="registry-server" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.565535 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="a97ed1ff-657f-4bde-943b-78caf9d07f92" containerName="registry-server" Jan 29 16:16:26 crc kubenswrapper[4714]: E0129 16:16:26.565547 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae853ba-61c9-439b-9dc9-21567075f18a" containerName="extract-utilities" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.565553 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae853ba-61c9-439b-9dc9-21567075f18a" containerName="extract-utilities" Jan 29 16:16:26 crc kubenswrapper[4714]: E0129 16:16:26.565563 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05e7c79-7d66-4453-aedb-f240784ff294" containerName="registry-server" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.565570 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05e7c79-7d66-4453-aedb-f240784ff294" containerName="registry-server" Jan 29 16:16:26 crc kubenswrapper[4714]: E0129 16:16:26.565579 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae853ba-61c9-439b-9dc9-21567075f18a" containerName="registry-server" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.565587 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae853ba-61c9-439b-9dc9-21567075f18a" containerName="registry-server" Jan 29 16:16:26 crc kubenswrapper[4714]: E0129 16:16:26.565597 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05e7c79-7d66-4453-aedb-f240784ff294" containerName="extract-content" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.565603 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05e7c79-7d66-4453-aedb-f240784ff294" containerName="extract-content" Jan 29 16:16:26 crc kubenswrapper[4714]: E0129 16:16:26.565611 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae853ba-61c9-439b-9dc9-21567075f18a" containerName="extract-content" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.565617 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae853ba-61c9-439b-9dc9-21567075f18a" containerName="extract-content" Jan 29 16:16:26 crc kubenswrapper[4714]: E0129 16:16:26.565623 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" containerName="registry-server" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.565629 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" containerName="registry-server" Jan 29 16:16:26 crc kubenswrapper[4714]: E0129 16:16:26.565637 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05e7c79-7d66-4453-aedb-f240784ff294" containerName="extract-utilities" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.565642 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05e7c79-7d66-4453-aedb-f240784ff294" containerName="extract-utilities" Jan 29 16:16:26 crc kubenswrapper[4714]: E0129 16:16:26.565652 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" containerName="extract-content" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.565657 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" containerName="extract-content" Jan 29 16:16:26 crc kubenswrapper[4714]: E0129 16:16:26.565665 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a97ed1ff-657f-4bde-943b-78caf9d07f92" containerName="extract-utilities" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.565671 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="a97ed1ff-657f-4bde-943b-78caf9d07f92" containerName="extract-utilities" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.565915 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="80515d06-c09e-4c9d-a90f-43cc84edf4c9" containerName="marketplace-operator" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.565939 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="d05e7c79-7d66-4453-aedb-f240784ff294" containerName="registry-server" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.565949 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="11a30de8-b234-47b4-8fd0-44f0c428be78" containerName="registry-server" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.565956 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="eae853ba-61c9-439b-9dc9-21567075f18a" containerName="registry-server" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.565965 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="a97ed1ff-657f-4bde-943b-78caf9d07f92" containerName="registry-server" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.566655 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6gkpz" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.569568 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.577042 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6gkpz"] Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.714862 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04dba3a0-a89b-48c5-97ef-e5660d1ae7bb-utilities\") pod \"redhat-marketplace-6gkpz\" (UID: \"04dba3a0-a89b-48c5-97ef-e5660d1ae7bb\") " pod="openshift-marketplace/redhat-marketplace-6gkpz" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.714910 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trwfb\" (UniqueName: \"kubernetes.io/projected/04dba3a0-a89b-48c5-97ef-e5660d1ae7bb-kube-api-access-trwfb\") pod \"redhat-marketplace-6gkpz\" (UID: \"04dba3a0-a89b-48c5-97ef-e5660d1ae7bb\") " pod="openshift-marketplace/redhat-marketplace-6gkpz" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.714967 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04dba3a0-a89b-48c5-97ef-e5660d1ae7bb-catalog-content\") pod \"redhat-marketplace-6gkpz\" (UID: \"04dba3a0-a89b-48c5-97ef-e5660d1ae7bb\") " pod="openshift-marketplace/redhat-marketplace-6gkpz" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.764213 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-knxc8"] Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.766571 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-knxc8" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.768196 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.808808 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-knxc8"] Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.816645 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04dba3a0-a89b-48c5-97ef-e5660d1ae7bb-utilities\") pod \"redhat-marketplace-6gkpz\" (UID: \"04dba3a0-a89b-48c5-97ef-e5660d1ae7bb\") " pod="openshift-marketplace/redhat-marketplace-6gkpz" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.816700 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trwfb\" (UniqueName: \"kubernetes.io/projected/04dba3a0-a89b-48c5-97ef-e5660d1ae7bb-kube-api-access-trwfb\") pod \"redhat-marketplace-6gkpz\" (UID: \"04dba3a0-a89b-48c5-97ef-e5660d1ae7bb\") " pod="openshift-marketplace/redhat-marketplace-6gkpz" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.816740 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04dba3a0-a89b-48c5-97ef-e5660d1ae7bb-catalog-content\") pod \"redhat-marketplace-6gkpz\" (UID: \"04dba3a0-a89b-48c5-97ef-e5660d1ae7bb\") " pod="openshift-marketplace/redhat-marketplace-6gkpz" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.817299 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04dba3a0-a89b-48c5-97ef-e5660d1ae7bb-utilities\") pod \"redhat-marketplace-6gkpz\" (UID: \"04dba3a0-a89b-48c5-97ef-e5660d1ae7bb\") " pod="openshift-marketplace/redhat-marketplace-6gkpz" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.817324 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04dba3a0-a89b-48c5-97ef-e5660d1ae7bb-catalog-content\") pod \"redhat-marketplace-6gkpz\" (UID: \"04dba3a0-a89b-48c5-97ef-e5660d1ae7bb\") " pod="openshift-marketplace/redhat-marketplace-6gkpz" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.834233 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trwfb\" (UniqueName: \"kubernetes.io/projected/04dba3a0-a89b-48c5-97ef-e5660d1ae7bb-kube-api-access-trwfb\") pod \"redhat-marketplace-6gkpz\" (UID: \"04dba3a0-a89b-48c5-97ef-e5660d1ae7bb\") " pod="openshift-marketplace/redhat-marketplace-6gkpz" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.916595 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6gkpz" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.917616 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de6c9fbd-8657-4434-bff5-468276791466-catalog-content\") pod \"redhat-operators-knxc8\" (UID: \"de6c9fbd-8657-4434-bff5-468276791466\") " pod="openshift-marketplace/redhat-operators-knxc8" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.917675 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de6c9fbd-8657-4434-bff5-468276791466-utilities\") pod \"redhat-operators-knxc8\" (UID: \"de6c9fbd-8657-4434-bff5-468276791466\") " pod="openshift-marketplace/redhat-operators-knxc8" Jan 29 16:16:26 crc kubenswrapper[4714]: I0129 16:16:26.917728 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75mht\" (UniqueName: \"kubernetes.io/projected/de6c9fbd-8657-4434-bff5-468276791466-kube-api-access-75mht\") pod \"redhat-operators-knxc8\" (UID: \"de6c9fbd-8657-4434-bff5-468276791466\") " pod="openshift-marketplace/redhat-operators-knxc8" Jan 29 16:16:27 crc kubenswrapper[4714]: I0129 16:16:27.018951 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de6c9fbd-8657-4434-bff5-468276791466-utilities\") pod \"redhat-operators-knxc8\" (UID: \"de6c9fbd-8657-4434-bff5-468276791466\") " pod="openshift-marketplace/redhat-operators-knxc8" Jan 29 16:16:27 crc kubenswrapper[4714]: I0129 16:16:27.019254 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75mht\" (UniqueName: \"kubernetes.io/projected/de6c9fbd-8657-4434-bff5-468276791466-kube-api-access-75mht\") pod \"redhat-operators-knxc8\" (UID: \"de6c9fbd-8657-4434-bff5-468276791466\") " pod="openshift-marketplace/redhat-operators-knxc8" Jan 29 16:16:27 crc kubenswrapper[4714]: I0129 16:16:27.019330 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de6c9fbd-8657-4434-bff5-468276791466-catalog-content\") pod \"redhat-operators-knxc8\" (UID: \"de6c9fbd-8657-4434-bff5-468276791466\") " pod="openshift-marketplace/redhat-operators-knxc8" Jan 29 16:16:27 crc kubenswrapper[4714]: I0129 16:16:27.019553 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de6c9fbd-8657-4434-bff5-468276791466-utilities\") pod \"redhat-operators-knxc8\" (UID: \"de6c9fbd-8657-4434-bff5-468276791466\") " pod="openshift-marketplace/redhat-operators-knxc8" Jan 29 16:16:27 crc kubenswrapper[4714]: I0129 16:16:27.019754 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de6c9fbd-8657-4434-bff5-468276791466-catalog-content\") pod \"redhat-operators-knxc8\" (UID: \"de6c9fbd-8657-4434-bff5-468276791466\") " pod="openshift-marketplace/redhat-operators-knxc8" Jan 29 16:16:27 crc kubenswrapper[4714]: I0129 16:16:27.037239 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75mht\" (UniqueName: \"kubernetes.io/projected/de6c9fbd-8657-4434-bff5-468276791466-kube-api-access-75mht\") pod \"redhat-operators-knxc8\" (UID: \"de6c9fbd-8657-4434-bff5-468276791466\") " pod="openshift-marketplace/redhat-operators-knxc8" Jan 29 16:16:27 crc kubenswrapper[4714]: I0129 16:16:27.089694 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-knxc8" Jan 29 16:16:27 crc kubenswrapper[4714]: I0129 16:16:27.128034 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6gkpz"] Jan 29 16:16:27 crc kubenswrapper[4714]: W0129 16:16:27.132607 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04dba3a0_a89b_48c5_97ef_e5660d1ae7bb.slice/crio-b49151a6cd20c5c170981578deb8f14550600f3f56ff6e1d1159c6320fbd9d56 WatchSource:0}: Error finding container b49151a6cd20c5c170981578deb8f14550600f3f56ff6e1d1159c6320fbd9d56: Status 404 returned error can't find the container with id b49151a6cd20c5c170981578deb8f14550600f3f56ff6e1d1159c6320fbd9d56 Jan 29 16:16:27 crc kubenswrapper[4714]: I0129 16:16:27.477512 4714 generic.go:334] "Generic (PLEG): container finished" podID="04dba3a0-a89b-48c5-97ef-e5660d1ae7bb" containerID="6787483294efed802480404e6f60af4b4a8206cdea6bbd7f4a0e458ef6b4913c" exitCode=0 Jan 29 16:16:27 crc kubenswrapper[4714]: I0129 16:16:27.479174 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6gkpz" event={"ID":"04dba3a0-a89b-48c5-97ef-e5660d1ae7bb","Type":"ContainerDied","Data":"6787483294efed802480404e6f60af4b4a8206cdea6bbd7f4a0e458ef6b4913c"} Jan 29 16:16:27 crc kubenswrapper[4714]: I0129 16:16:27.479204 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6gkpz" event={"ID":"04dba3a0-a89b-48c5-97ef-e5660d1ae7bb","Type":"ContainerStarted","Data":"b49151a6cd20c5c170981578deb8f14550600f3f56ff6e1d1159c6320fbd9d56"} Jan 29 16:16:27 crc kubenswrapper[4714]: I0129 16:16:27.529659 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-knxc8"] Jan 29 16:16:27 crc kubenswrapper[4714]: I0129 16:16:27.844830 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:16:27 crc kubenswrapper[4714]: I0129 16:16:27.845262 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:16:28 crc kubenswrapper[4714]: I0129 16:16:28.483724 4714 generic.go:334] "Generic (PLEG): container finished" podID="de6c9fbd-8657-4434-bff5-468276791466" containerID="f21d6daee17eb6d985ca58354495191425d749da3fa45be5996908ac6f08780f" exitCode=0 Jan 29 16:16:28 crc kubenswrapper[4714]: I0129 16:16:28.483828 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knxc8" event={"ID":"de6c9fbd-8657-4434-bff5-468276791466","Type":"ContainerDied","Data":"f21d6daee17eb6d985ca58354495191425d749da3fa45be5996908ac6f08780f"} Jan 29 16:16:28 crc kubenswrapper[4714]: I0129 16:16:28.483871 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knxc8" event={"ID":"de6c9fbd-8657-4434-bff5-468276791466","Type":"ContainerStarted","Data":"29b015d32919088e769aeed20bae467711af5bb04181697f47d676a18c6bd238"} Jan 29 16:16:28 crc kubenswrapper[4714]: I0129 16:16:28.486123 4714 generic.go:334] "Generic (PLEG): container finished" podID="04dba3a0-a89b-48c5-97ef-e5660d1ae7bb" containerID="d0f6caf51e1515d83dc6f892c64af223c43e57b8b5ed10f67a693edcb05bd71e" exitCode=0 Jan 29 16:16:28 crc kubenswrapper[4714]: I0129 16:16:28.486164 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6gkpz" event={"ID":"04dba3a0-a89b-48c5-97ef-e5660d1ae7bb","Type":"ContainerDied","Data":"d0f6caf51e1515d83dc6f892c64af223c43e57b8b5ed10f67a693edcb05bd71e"} Jan 29 16:16:28 crc kubenswrapper[4714]: I0129 16:16:28.963785 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-slsxz"] Jan 29 16:16:28 crc kubenswrapper[4714]: I0129 16:16:28.965961 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-slsxz" Jan 29 16:16:28 crc kubenswrapper[4714]: I0129 16:16:28.969975 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 29 16:16:28 crc kubenswrapper[4714]: I0129 16:16:28.973688 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-slsxz"] Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.047689 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16cb244c-6c63-47e6-a312-ba33ab4d4899-catalog-content\") pod \"certified-operators-slsxz\" (UID: \"16cb244c-6c63-47e6-a312-ba33ab4d4899\") " pod="openshift-marketplace/certified-operators-slsxz" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.047728 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsgc7\" (UniqueName: \"kubernetes.io/projected/16cb244c-6c63-47e6-a312-ba33ab4d4899-kube-api-access-jsgc7\") pod \"certified-operators-slsxz\" (UID: \"16cb244c-6c63-47e6-a312-ba33ab4d4899\") " pod="openshift-marketplace/certified-operators-slsxz" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.047760 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16cb244c-6c63-47e6-a312-ba33ab4d4899-utilities\") pod \"certified-operators-slsxz\" (UID: \"16cb244c-6c63-47e6-a312-ba33ab4d4899\") " pod="openshift-marketplace/certified-operators-slsxz" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.149266 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16cb244c-6c63-47e6-a312-ba33ab4d4899-catalog-content\") pod \"certified-operators-slsxz\" (UID: \"16cb244c-6c63-47e6-a312-ba33ab4d4899\") " pod="openshift-marketplace/certified-operators-slsxz" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.149625 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsgc7\" (UniqueName: \"kubernetes.io/projected/16cb244c-6c63-47e6-a312-ba33ab4d4899-kube-api-access-jsgc7\") pod \"certified-operators-slsxz\" (UID: \"16cb244c-6c63-47e6-a312-ba33ab4d4899\") " pod="openshift-marketplace/certified-operators-slsxz" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.149674 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16cb244c-6c63-47e6-a312-ba33ab4d4899-utilities\") pod \"certified-operators-slsxz\" (UID: \"16cb244c-6c63-47e6-a312-ba33ab4d4899\") " pod="openshift-marketplace/certified-operators-slsxz" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.150369 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16cb244c-6c63-47e6-a312-ba33ab4d4899-utilities\") pod \"certified-operators-slsxz\" (UID: \"16cb244c-6c63-47e6-a312-ba33ab4d4899\") " pod="openshift-marketplace/certified-operators-slsxz" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.151269 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16cb244c-6c63-47e6-a312-ba33ab4d4899-catalog-content\") pod \"certified-operators-slsxz\" (UID: \"16cb244c-6c63-47e6-a312-ba33ab4d4899\") " pod="openshift-marketplace/certified-operators-slsxz" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.168401 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ndx6p"] Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.170011 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ndx6p" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.172294 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.178807 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsgc7\" (UniqueName: \"kubernetes.io/projected/16cb244c-6c63-47e6-a312-ba33ab4d4899-kube-api-access-jsgc7\") pod \"certified-operators-slsxz\" (UID: \"16cb244c-6c63-47e6-a312-ba33ab4d4899\") " pod="openshift-marketplace/certified-operators-slsxz" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.190980 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ndx6p"] Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.251065 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnd9z\" (UniqueName: \"kubernetes.io/projected/ca655e22-8f97-4e9e-b115-734ae1af7d50-kube-api-access-qnd9z\") pod \"community-operators-ndx6p\" (UID: \"ca655e22-8f97-4e9e-b115-734ae1af7d50\") " pod="openshift-marketplace/community-operators-ndx6p" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.251116 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca655e22-8f97-4e9e-b115-734ae1af7d50-catalog-content\") pod \"community-operators-ndx6p\" (UID: \"ca655e22-8f97-4e9e-b115-734ae1af7d50\") " pod="openshift-marketplace/community-operators-ndx6p" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.251140 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca655e22-8f97-4e9e-b115-734ae1af7d50-utilities\") pod \"community-operators-ndx6p\" (UID: \"ca655e22-8f97-4e9e-b115-734ae1af7d50\") " pod="openshift-marketplace/community-operators-ndx6p" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.322050 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-slsxz" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.352014 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnd9z\" (UniqueName: \"kubernetes.io/projected/ca655e22-8f97-4e9e-b115-734ae1af7d50-kube-api-access-qnd9z\") pod \"community-operators-ndx6p\" (UID: \"ca655e22-8f97-4e9e-b115-734ae1af7d50\") " pod="openshift-marketplace/community-operators-ndx6p" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.352078 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca655e22-8f97-4e9e-b115-734ae1af7d50-catalog-content\") pod \"community-operators-ndx6p\" (UID: \"ca655e22-8f97-4e9e-b115-734ae1af7d50\") " pod="openshift-marketplace/community-operators-ndx6p" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.352104 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca655e22-8f97-4e9e-b115-734ae1af7d50-utilities\") pod \"community-operators-ndx6p\" (UID: \"ca655e22-8f97-4e9e-b115-734ae1af7d50\") " pod="openshift-marketplace/community-operators-ndx6p" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.352487 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca655e22-8f97-4e9e-b115-734ae1af7d50-utilities\") pod \"community-operators-ndx6p\" (UID: \"ca655e22-8f97-4e9e-b115-734ae1af7d50\") " pod="openshift-marketplace/community-operators-ndx6p" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.353466 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca655e22-8f97-4e9e-b115-734ae1af7d50-catalog-content\") pod \"community-operators-ndx6p\" (UID: \"ca655e22-8f97-4e9e-b115-734ae1af7d50\") " pod="openshift-marketplace/community-operators-ndx6p" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.372744 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnd9z\" (UniqueName: \"kubernetes.io/projected/ca655e22-8f97-4e9e-b115-734ae1af7d50-kube-api-access-qnd9z\") pod \"community-operators-ndx6p\" (UID: \"ca655e22-8f97-4e9e-b115-734ae1af7d50\") " pod="openshift-marketplace/community-operators-ndx6p" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.502433 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6gkpz" event={"ID":"04dba3a0-a89b-48c5-97ef-e5660d1ae7bb","Type":"ContainerStarted","Data":"35fc0ccefdc74ae4d4ac8a83e158c0a04ec3c4c8190a9cb3d91d2dd6bb5fe65e"} Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.512052 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ndx6p" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.526569 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6gkpz" podStartSLOduration=2.09524823 podStartE2EDuration="3.526552168s" podCreationTimestamp="2026-01-29 16:16:26 +0000 UTC" firstStartedPulling="2026-01-29 16:16:27.480421539 +0000 UTC m=+394.000922659" lastFinishedPulling="2026-01-29 16:16:28.911725477 +0000 UTC m=+395.432226597" observedRunningTime="2026-01-29 16:16:29.52419837 +0000 UTC m=+396.044699490" watchObservedRunningTime="2026-01-29 16:16:29.526552168 +0000 UTC m=+396.047053288" Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.536488 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-slsxz"] Jan 29 16:16:29 crc kubenswrapper[4714]: W0129 16:16:29.545649 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16cb244c_6c63_47e6_a312_ba33ab4d4899.slice/crio-884b32749714c8c7d5a95398be26c6c16b6bc585a925ce02a67b3f4c885cba1c WatchSource:0}: Error finding container 884b32749714c8c7d5a95398be26c6c16b6bc585a925ce02a67b3f4c885cba1c: Status 404 returned error can't find the container with id 884b32749714c8c7d5a95398be26c6c16b6bc585a925ce02a67b3f4c885cba1c Jan 29 16:16:29 crc kubenswrapper[4714]: I0129 16:16:29.916848 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ndx6p"] Jan 29 16:16:29 crc kubenswrapper[4714]: W0129 16:16:29.923094 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca655e22_8f97_4e9e_b115_734ae1af7d50.slice/crio-756fbfa36039add55873032181de3423d6d3224932100855a9a3c1eb47efd1fc WatchSource:0}: Error finding container 756fbfa36039add55873032181de3423d6d3224932100855a9a3c1eb47efd1fc: Status 404 returned error can't find the container with id 756fbfa36039add55873032181de3423d6d3224932100855a9a3c1eb47efd1fc Jan 29 16:16:30 crc kubenswrapper[4714]: I0129 16:16:30.053897 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-xrw9s" Jan 29 16:16:30 crc kubenswrapper[4714]: I0129 16:16:30.119949 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gnjmm"] Jan 29 16:16:30 crc kubenswrapper[4714]: I0129 16:16:30.509281 4714 generic.go:334] "Generic (PLEG): container finished" podID="ca655e22-8f97-4e9e-b115-734ae1af7d50" containerID="9a954353aef73625a94aef15b9031dd32efca16dc8803b36934b6c950db5a96c" exitCode=0 Jan 29 16:16:30 crc kubenswrapper[4714]: I0129 16:16:30.509331 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndx6p" event={"ID":"ca655e22-8f97-4e9e-b115-734ae1af7d50","Type":"ContainerDied","Data":"9a954353aef73625a94aef15b9031dd32efca16dc8803b36934b6c950db5a96c"} Jan 29 16:16:30 crc kubenswrapper[4714]: I0129 16:16:30.509377 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndx6p" event={"ID":"ca655e22-8f97-4e9e-b115-734ae1af7d50","Type":"ContainerStarted","Data":"756fbfa36039add55873032181de3423d6d3224932100855a9a3c1eb47efd1fc"} Jan 29 16:16:30 crc kubenswrapper[4714]: I0129 16:16:30.511194 4714 generic.go:334] "Generic (PLEG): container finished" podID="16cb244c-6c63-47e6-a312-ba33ab4d4899" containerID="62d59d1936ff31d8ea6d4c8d7ff82c3652ecb0ea4515c3877dfee41f17b9ee07" exitCode=0 Jan 29 16:16:30 crc kubenswrapper[4714]: I0129 16:16:30.511267 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slsxz" event={"ID":"16cb244c-6c63-47e6-a312-ba33ab4d4899","Type":"ContainerDied","Data":"62d59d1936ff31d8ea6d4c8d7ff82c3652ecb0ea4515c3877dfee41f17b9ee07"} Jan 29 16:16:30 crc kubenswrapper[4714]: I0129 16:16:30.511300 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slsxz" event={"ID":"16cb244c-6c63-47e6-a312-ba33ab4d4899","Type":"ContainerStarted","Data":"884b32749714c8c7d5a95398be26c6c16b6bc585a925ce02a67b3f4c885cba1c"} Jan 29 16:16:30 crc kubenswrapper[4714]: I0129 16:16:30.514266 4714 generic.go:334] "Generic (PLEG): container finished" podID="de6c9fbd-8657-4434-bff5-468276791466" containerID="f26a96b15ed747c0a4867efe91365ae7ff513a3b7a8ae9f0e7807a4e2f772cdc" exitCode=0 Jan 29 16:16:30 crc kubenswrapper[4714]: I0129 16:16:30.514319 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knxc8" event={"ID":"de6c9fbd-8657-4434-bff5-468276791466","Type":"ContainerDied","Data":"f26a96b15ed747c0a4867efe91365ae7ff513a3b7a8ae9f0e7807a4e2f772cdc"} Jan 29 16:16:31 crc kubenswrapper[4714]: I0129 16:16:31.521186 4714 generic.go:334] "Generic (PLEG): container finished" podID="ca655e22-8f97-4e9e-b115-734ae1af7d50" containerID="7c8138352f6edb954a09ba035a20f9a65c4b7052c054970be09c397de55273a1" exitCode=0 Jan 29 16:16:31 crc kubenswrapper[4714]: I0129 16:16:31.521705 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndx6p" event={"ID":"ca655e22-8f97-4e9e-b115-734ae1af7d50","Type":"ContainerDied","Data":"7c8138352f6edb954a09ba035a20f9a65c4b7052c054970be09c397de55273a1"} Jan 29 16:16:31 crc kubenswrapper[4714]: I0129 16:16:31.524339 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slsxz" event={"ID":"16cb244c-6c63-47e6-a312-ba33ab4d4899","Type":"ContainerStarted","Data":"3bc5b72a0a9952ffe2ca554b13a97237dd7b15673e19bc821cf21f1143b8ae85"} Jan 29 16:16:31 crc kubenswrapper[4714]: I0129 16:16:31.528462 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knxc8" event={"ID":"de6c9fbd-8657-4434-bff5-468276791466","Type":"ContainerStarted","Data":"3649c308dddf3d4460709e70403d1db1343f0378f932455c872f88fc66376358"} Jan 29 16:16:31 crc kubenswrapper[4714]: I0129 16:16:31.561511 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-knxc8" podStartSLOduration=3.064075607 podStartE2EDuration="5.56148611s" podCreationTimestamp="2026-01-29 16:16:26 +0000 UTC" firstStartedPulling="2026-01-29 16:16:28.486003545 +0000 UTC m=+395.006504665" lastFinishedPulling="2026-01-29 16:16:30.983414048 +0000 UTC m=+397.503915168" observedRunningTime="2026-01-29 16:16:31.560590964 +0000 UTC m=+398.081092104" watchObservedRunningTime="2026-01-29 16:16:31.56148611 +0000 UTC m=+398.081987230" Jan 29 16:16:32 crc kubenswrapper[4714]: I0129 16:16:32.535489 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndx6p" event={"ID":"ca655e22-8f97-4e9e-b115-734ae1af7d50","Type":"ContainerStarted","Data":"a9c4299a993eca4f0aa5c07550820a1ed2d795a2558e78a49fa48d90534cf65d"} Jan 29 16:16:32 crc kubenswrapper[4714]: I0129 16:16:32.537443 4714 generic.go:334] "Generic (PLEG): container finished" podID="16cb244c-6c63-47e6-a312-ba33ab4d4899" containerID="3bc5b72a0a9952ffe2ca554b13a97237dd7b15673e19bc821cf21f1143b8ae85" exitCode=0 Jan 29 16:16:32 crc kubenswrapper[4714]: I0129 16:16:32.537517 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slsxz" event={"ID":"16cb244c-6c63-47e6-a312-ba33ab4d4899","Type":"ContainerDied","Data":"3bc5b72a0a9952ffe2ca554b13a97237dd7b15673e19bc821cf21f1143b8ae85"} Jan 29 16:16:32 crc kubenswrapper[4714]: I0129 16:16:32.537580 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slsxz" event={"ID":"16cb244c-6c63-47e6-a312-ba33ab4d4899","Type":"ContainerStarted","Data":"8cebb6abc380ded14169355538a0129fc12d3b96fafd58e18998a7ab9b5ba7ab"} Jan 29 16:16:32 crc kubenswrapper[4714]: I0129 16:16:32.553161 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ndx6p" podStartSLOduration=2.144912214 podStartE2EDuration="3.553140611s" podCreationTimestamp="2026-01-29 16:16:29 +0000 UTC" firstStartedPulling="2026-01-29 16:16:30.510974837 +0000 UTC m=+397.031475957" lastFinishedPulling="2026-01-29 16:16:31.919203224 +0000 UTC m=+398.439704354" observedRunningTime="2026-01-29 16:16:32.550543635 +0000 UTC m=+399.071044755" watchObservedRunningTime="2026-01-29 16:16:32.553140611 +0000 UTC m=+399.073641731" Jan 29 16:16:32 crc kubenswrapper[4714]: I0129 16:16:32.577754 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-slsxz" podStartSLOduration=3.148554021 podStartE2EDuration="4.577731336s" podCreationTimestamp="2026-01-29 16:16:28 +0000 UTC" firstStartedPulling="2026-01-29 16:16:30.51277732 +0000 UTC m=+397.033278440" lastFinishedPulling="2026-01-29 16:16:31.941954595 +0000 UTC m=+398.462455755" observedRunningTime="2026-01-29 16:16:32.574538673 +0000 UTC m=+399.095039793" watchObservedRunningTime="2026-01-29 16:16:32.577731336 +0000 UTC m=+399.098232456" Jan 29 16:16:36 crc kubenswrapper[4714]: I0129 16:16:36.917555 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6gkpz" Jan 29 16:16:36 crc kubenswrapper[4714]: I0129 16:16:36.918239 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6gkpz" Jan 29 16:16:36 crc kubenswrapper[4714]: I0129 16:16:36.959472 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6gkpz" Jan 29 16:16:37 crc kubenswrapper[4714]: I0129 16:16:37.090524 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-knxc8" Jan 29 16:16:37 crc kubenswrapper[4714]: I0129 16:16:37.090590 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-knxc8" Jan 29 16:16:37 crc kubenswrapper[4714]: I0129 16:16:37.626222 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6gkpz" Jan 29 16:16:38 crc kubenswrapper[4714]: I0129 16:16:38.143942 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-knxc8" podUID="de6c9fbd-8657-4434-bff5-468276791466" containerName="registry-server" probeResult="failure" output=< Jan 29 16:16:38 crc kubenswrapper[4714]: timeout: failed to connect service ":50051" within 1s Jan 29 16:16:38 crc kubenswrapper[4714]: > Jan 29 16:16:39 crc kubenswrapper[4714]: I0129 16:16:39.322922 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-slsxz" Jan 29 16:16:39 crc kubenswrapper[4714]: I0129 16:16:39.323001 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-slsxz" Jan 29 16:16:39 crc kubenswrapper[4714]: I0129 16:16:39.364485 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-slsxz" Jan 29 16:16:39 crc kubenswrapper[4714]: I0129 16:16:39.512872 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ndx6p" Jan 29 16:16:39 crc kubenswrapper[4714]: I0129 16:16:39.512923 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ndx6p" Jan 29 16:16:39 crc kubenswrapper[4714]: I0129 16:16:39.562944 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ndx6p" Jan 29 16:16:39 crc kubenswrapper[4714]: I0129 16:16:39.618783 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ndx6p" Jan 29 16:16:39 crc kubenswrapper[4714]: I0129 16:16:39.619951 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-slsxz" Jan 29 16:16:47 crc kubenswrapper[4714]: I0129 16:16:47.153122 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-knxc8" Jan 29 16:16:47 crc kubenswrapper[4714]: I0129 16:16:47.217000 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-knxc8" Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.170112 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" podUID="48be8ad8-4c02-4bea-a143-449763b39d54" containerName="registry" containerID="cri-o://f2810d7fe0cd9711b0f9c4dd754cc2bb22760a85d91d92553df7afe7b8378eac" gracePeriod=30 Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.567553 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.660022 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/48be8ad8-4c02-4bea-a143-449763b39d54-trusted-ca\") pod \"48be8ad8-4c02-4bea-a143-449763b39d54\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.660196 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/48be8ad8-4c02-4bea-a143-449763b39d54-registry-tls\") pod \"48be8ad8-4c02-4bea-a143-449763b39d54\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.660350 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/48be8ad8-4c02-4bea-a143-449763b39d54-registry-certificates\") pod \"48be8ad8-4c02-4bea-a143-449763b39d54\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.660424 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/48be8ad8-4c02-4bea-a143-449763b39d54-ca-trust-extracted\") pod \"48be8ad8-4c02-4bea-a143-449763b39d54\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.660811 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"48be8ad8-4c02-4bea-a143-449763b39d54\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.660913 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/48be8ad8-4c02-4bea-a143-449763b39d54-installation-pull-secrets\") pod \"48be8ad8-4c02-4bea-a143-449763b39d54\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.661050 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/48be8ad8-4c02-4bea-a143-449763b39d54-bound-sa-token\") pod \"48be8ad8-4c02-4bea-a143-449763b39d54\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.661346 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wh5km\" (UniqueName: \"kubernetes.io/projected/48be8ad8-4c02-4bea-a143-449763b39d54-kube-api-access-wh5km\") pod \"48be8ad8-4c02-4bea-a143-449763b39d54\" (UID: \"48be8ad8-4c02-4bea-a143-449763b39d54\") " Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.663290 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48be8ad8-4c02-4bea-a143-449763b39d54-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "48be8ad8-4c02-4bea-a143-449763b39d54" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.663389 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48be8ad8-4c02-4bea-a143-449763b39d54-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "48be8ad8-4c02-4bea-a143-449763b39d54" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.664095 4714 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/48be8ad8-4c02-4bea-a143-449763b39d54-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.664131 4714 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/48be8ad8-4c02-4bea-a143-449763b39d54-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.667017 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48be8ad8-4c02-4bea-a143-449763b39d54-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "48be8ad8-4c02-4bea-a143-449763b39d54" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.670246 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48be8ad8-4c02-4bea-a143-449763b39d54-kube-api-access-wh5km" (OuterVolumeSpecName: "kube-api-access-wh5km") pod "48be8ad8-4c02-4bea-a143-449763b39d54" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54"). InnerVolumeSpecName "kube-api-access-wh5km". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.674066 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48be8ad8-4c02-4bea-a143-449763b39d54-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "48be8ad8-4c02-4bea-a143-449763b39d54" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.674188 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48be8ad8-4c02-4bea-a143-449763b39d54-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "48be8ad8-4c02-4bea-a143-449763b39d54" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.675060 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "48be8ad8-4c02-4bea-a143-449763b39d54" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.675767 4714 generic.go:334] "Generic (PLEG): container finished" podID="48be8ad8-4c02-4bea-a143-449763b39d54" containerID="f2810d7fe0cd9711b0f9c4dd754cc2bb22760a85d91d92553df7afe7b8378eac" exitCode=0 Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.675815 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" event={"ID":"48be8ad8-4c02-4bea-a143-449763b39d54","Type":"ContainerDied","Data":"f2810d7fe0cd9711b0f9c4dd754cc2bb22760a85d91d92553df7afe7b8378eac"} Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.675849 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" event={"ID":"48be8ad8-4c02-4bea-a143-449763b39d54","Type":"ContainerDied","Data":"815b16152db25222f3f6a5ff40233d8cdbe464e73d20d130a327746193531954"} Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.675872 4714 scope.go:117] "RemoveContainer" containerID="f2810d7fe0cd9711b0f9c4dd754cc2bb22760a85d91d92553df7afe7b8378eac" Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.676038 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gnjmm" Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.696237 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48be8ad8-4c02-4bea-a143-449763b39d54-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "48be8ad8-4c02-4bea-a143-449763b39d54" (UID: "48be8ad8-4c02-4bea-a143-449763b39d54"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.714424 4714 scope.go:117] "RemoveContainer" containerID="f2810d7fe0cd9711b0f9c4dd754cc2bb22760a85d91d92553df7afe7b8378eac" Jan 29 16:16:55 crc kubenswrapper[4714]: E0129 16:16:55.714999 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2810d7fe0cd9711b0f9c4dd754cc2bb22760a85d91d92553df7afe7b8378eac\": container with ID starting with f2810d7fe0cd9711b0f9c4dd754cc2bb22760a85d91d92553df7afe7b8378eac not found: ID does not exist" containerID="f2810d7fe0cd9711b0f9c4dd754cc2bb22760a85d91d92553df7afe7b8378eac" Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.715034 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2810d7fe0cd9711b0f9c4dd754cc2bb22760a85d91d92553df7afe7b8378eac"} err="failed to get container status \"f2810d7fe0cd9711b0f9c4dd754cc2bb22760a85d91d92553df7afe7b8378eac\": rpc error: code = NotFound desc = could not find container \"f2810d7fe0cd9711b0f9c4dd754cc2bb22760a85d91d92553df7afe7b8378eac\": container with ID starting with f2810d7fe0cd9711b0f9c4dd754cc2bb22760a85d91d92553df7afe7b8378eac not found: ID does not exist" Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.765699 4714 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/48be8ad8-4c02-4bea-a143-449763b39d54-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.765760 4714 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/48be8ad8-4c02-4bea-a143-449763b39d54-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.765781 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wh5km\" (UniqueName: \"kubernetes.io/projected/48be8ad8-4c02-4bea-a143-449763b39d54-kube-api-access-wh5km\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.765798 4714 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/48be8ad8-4c02-4bea-a143-449763b39d54-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:55 crc kubenswrapper[4714]: I0129 16:16:55.765817 4714 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/48be8ad8-4c02-4bea-a143-449763b39d54-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 29 16:16:56 crc kubenswrapper[4714]: I0129 16:16:56.036190 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gnjmm"] Jan 29 16:16:56 crc kubenswrapper[4714]: I0129 16:16:56.042903 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gnjmm"] Jan 29 16:16:56 crc kubenswrapper[4714]: I0129 16:16:56.195530 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48be8ad8-4c02-4bea-a143-449763b39d54" path="/var/lib/kubelet/pods/48be8ad8-4c02-4bea-a143-449763b39d54/volumes" Jan 29 16:16:57 crc kubenswrapper[4714]: I0129 16:16:57.853070 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:16:57 crc kubenswrapper[4714]: I0129 16:16:57.853169 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:16:57 crc kubenswrapper[4714]: I0129 16:16:57.853239 4714 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:16:57 crc kubenswrapper[4714]: I0129 16:16:57.854139 4714 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"32d59d2a4eb095db60ac3365411265035b47b1f01d164e950c86daa5aecb2792"} pod="openshift-machine-config-operator/machine-config-daemon-ppngk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 16:16:57 crc kubenswrapper[4714]: I0129 16:16:57.854247 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" containerID="cri-o://32d59d2a4eb095db60ac3365411265035b47b1f01d164e950c86daa5aecb2792" gracePeriod=600 Jan 29 16:16:58 crc kubenswrapper[4714]: I0129 16:16:58.708020 4714 generic.go:334] "Generic (PLEG): container finished" podID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerID="32d59d2a4eb095db60ac3365411265035b47b1f01d164e950c86daa5aecb2792" exitCode=0 Jan 29 16:16:58 crc kubenswrapper[4714]: I0129 16:16:58.708104 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" event={"ID":"c8c765f3-89eb-4077-8829-03e86eb0c90c","Type":"ContainerDied","Data":"32d59d2a4eb095db60ac3365411265035b47b1f01d164e950c86daa5aecb2792"} Jan 29 16:16:58 crc kubenswrapper[4714]: I0129 16:16:58.708840 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" event={"ID":"c8c765f3-89eb-4077-8829-03e86eb0c90c","Type":"ContainerStarted","Data":"aeda778ca6de188bfb9f09408c5d355e6f8d4366d5f9ebe7bfd9f2e4dea2a0e4"} Jan 29 16:16:58 crc kubenswrapper[4714]: I0129 16:16:58.708907 4714 scope.go:117] "RemoveContainer" containerID="27cd83775817b7c8fd45f33899dd9a718067500e7a4853c38451161035fd33e5" Jan 29 16:19:27 crc kubenswrapper[4714]: I0129 16:19:27.844626 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:19:27 crc kubenswrapper[4714]: I0129 16:19:27.845174 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:19:57 crc kubenswrapper[4714]: I0129 16:19:57.844058 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:19:57 crc kubenswrapper[4714]: I0129 16:19:57.844882 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:20:27 crc kubenswrapper[4714]: I0129 16:20:27.844453 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:20:27 crc kubenswrapper[4714]: I0129 16:20:27.844904 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:20:27 crc kubenswrapper[4714]: I0129 16:20:27.844978 4714 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:20:27 crc kubenswrapper[4714]: I0129 16:20:27.845585 4714 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aeda778ca6de188bfb9f09408c5d355e6f8d4366d5f9ebe7bfd9f2e4dea2a0e4"} pod="openshift-machine-config-operator/machine-config-daemon-ppngk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 16:20:27 crc kubenswrapper[4714]: I0129 16:20:27.845649 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" containerID="cri-o://aeda778ca6de188bfb9f09408c5d355e6f8d4366d5f9ebe7bfd9f2e4dea2a0e4" gracePeriod=600 Jan 29 16:20:28 crc kubenswrapper[4714]: I0129 16:20:28.341451 4714 generic.go:334] "Generic (PLEG): container finished" podID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerID="aeda778ca6de188bfb9f09408c5d355e6f8d4366d5f9ebe7bfd9f2e4dea2a0e4" exitCode=0 Jan 29 16:20:28 crc kubenswrapper[4714]: I0129 16:20:28.341535 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" event={"ID":"c8c765f3-89eb-4077-8829-03e86eb0c90c","Type":"ContainerDied","Data":"aeda778ca6de188bfb9f09408c5d355e6f8d4366d5f9ebe7bfd9f2e4dea2a0e4"} Jan 29 16:20:28 crc kubenswrapper[4714]: I0129 16:20:28.342061 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" event={"ID":"c8c765f3-89eb-4077-8829-03e86eb0c90c","Type":"ContainerStarted","Data":"434181b332ad91829c9ca3b07c475cac7d3c8b013492e90ce07fd88776d24efa"} Jan 29 16:20:28 crc kubenswrapper[4714]: I0129 16:20:28.342098 4714 scope.go:117] "RemoveContainer" containerID="32d59d2a4eb095db60ac3365411265035b47b1f01d164e950c86daa5aecb2792" Jan 29 16:21:48 crc kubenswrapper[4714]: I0129 16:21:48.800121 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-sbnkt"] Jan 29 16:21:48 crc kubenswrapper[4714]: I0129 16:21:48.801221 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovn-controller" containerID="cri-o://7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016" gracePeriod=30 Jan 29 16:21:48 crc kubenswrapper[4714]: I0129 16:21:48.801612 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="sbdb" containerID="cri-o://662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf" gracePeriod=30 Jan 29 16:21:48 crc kubenswrapper[4714]: I0129 16:21:48.801657 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="nbdb" containerID="cri-o://5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7" gracePeriod=30 Jan 29 16:21:48 crc kubenswrapper[4714]: I0129 16:21:48.801695 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="northd" containerID="cri-o://e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4" gracePeriod=30 Jan 29 16:21:48 crc kubenswrapper[4714]: I0129 16:21:48.801750 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816" gracePeriod=30 Jan 29 16:21:48 crc kubenswrapper[4714]: I0129 16:21:48.801828 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="kube-rbac-proxy-node" containerID="cri-o://b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc" gracePeriod=30 Jan 29 16:21:48 crc kubenswrapper[4714]: I0129 16:21:48.801870 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovn-acl-logging" containerID="cri-o://2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd" gracePeriod=30 Jan 29 16:21:48 crc kubenswrapper[4714]: I0129 16:21:48.836364 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovnkube-controller" containerID="cri-o://00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6" gracePeriod=30 Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.722074 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-sbnkt_04b20f02-6c1e-4082-8233-8f06bda63195/ovnkube-controller/3.log" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.725477 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-sbnkt_04b20f02-6c1e-4082-8233-8f06bda63195/ovn-acl-logging/0.log" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.726181 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-sbnkt_04b20f02-6c1e-4082-8233-8f06bda63195/ovn-controller/0.log" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.726919 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.795741 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-skzvq"] Jan 29 16:21:49 crc kubenswrapper[4714]: E0129 16:21:49.796003 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="kube-rbac-proxy-ovn-metrics" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796021 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="kube-rbac-proxy-ovn-metrics" Jan 29 16:21:49 crc kubenswrapper[4714]: E0129 16:21:49.796032 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48be8ad8-4c02-4bea-a143-449763b39d54" containerName="registry" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796041 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="48be8ad8-4c02-4bea-a143-449763b39d54" containerName="registry" Jan 29 16:21:49 crc kubenswrapper[4714]: E0129 16:21:49.796051 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="kubecfg-setup" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796058 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="kubecfg-setup" Jan 29 16:21:49 crc kubenswrapper[4714]: E0129 16:21:49.796066 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="sbdb" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796074 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="sbdb" Jan 29 16:21:49 crc kubenswrapper[4714]: E0129 16:21:49.796084 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="northd" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796091 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="northd" Jan 29 16:21:49 crc kubenswrapper[4714]: E0129 16:21:49.796103 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovnkube-controller" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796110 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovnkube-controller" Jan 29 16:21:49 crc kubenswrapper[4714]: E0129 16:21:49.796119 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="nbdb" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796128 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="nbdb" Jan 29 16:21:49 crc kubenswrapper[4714]: E0129 16:21:49.796136 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovnkube-controller" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796143 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovnkube-controller" Jan 29 16:21:49 crc kubenswrapper[4714]: E0129 16:21:49.796157 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="kube-rbac-proxy-node" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796165 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="kube-rbac-proxy-node" Jan 29 16:21:49 crc kubenswrapper[4714]: E0129 16:21:49.796174 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovnkube-controller" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796182 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovnkube-controller" Jan 29 16:21:49 crc kubenswrapper[4714]: E0129 16:21:49.796192 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovnkube-controller" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796198 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovnkube-controller" Jan 29 16:21:49 crc kubenswrapper[4714]: E0129 16:21:49.796206 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovn-acl-logging" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796212 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovn-acl-logging" Jan 29 16:21:49 crc kubenswrapper[4714]: E0129 16:21:49.796222 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovn-controller" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796228 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovn-controller" Jan 29 16:21:49 crc kubenswrapper[4714]: E0129 16:21:49.796236 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovnkube-controller" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796241 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovnkube-controller" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796320 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovnkube-controller" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796329 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="kube-rbac-proxy-ovn-metrics" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796337 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovnkube-controller" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796346 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="48be8ad8-4c02-4bea-a143-449763b39d54" containerName="registry" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796353 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="nbdb" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796360 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="kube-rbac-proxy-node" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796367 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovn-controller" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796375 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="sbdb" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796381 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovn-acl-logging" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796389 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="northd" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796396 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovnkube-controller" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796543 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovnkube-controller" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.796553 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" containerName="ovnkube-controller" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.798003 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886490 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-run-ovn\") pod \"04b20f02-6c1e-4082-8233-8f06bda63195\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886546 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-var-lib-cni-networks-ovn-kubernetes\") pod \"04b20f02-6c1e-4082-8233-8f06bda63195\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886558 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "04b20f02-6c1e-4082-8233-8f06bda63195" (UID: "04b20f02-6c1e-4082-8233-8f06bda63195"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886579 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-cni-netd\") pod \"04b20f02-6c1e-4082-8233-8f06bda63195\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886609 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "04b20f02-6c1e-4082-8233-8f06bda63195" (UID: "04b20f02-6c1e-4082-8233-8f06bda63195"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886610 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-var-lib-openvswitch\") pod \"04b20f02-6c1e-4082-8233-8f06bda63195\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886660 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-slash\") pod \"04b20f02-6c1e-4082-8233-8f06bda63195\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886665 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "04b20f02-6c1e-4082-8233-8f06bda63195" (UID: "04b20f02-6c1e-4082-8233-8f06bda63195"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886688 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-run-netns\") pod \"04b20f02-6c1e-4082-8233-8f06bda63195\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886714 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-run-ovn-kubernetes\") pod \"04b20f02-6c1e-4082-8233-8f06bda63195\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886767 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vrsm\" (UniqueName: \"kubernetes.io/projected/04b20f02-6c1e-4082-8233-8f06bda63195-kube-api-access-7vrsm\") pod \"04b20f02-6c1e-4082-8233-8f06bda63195\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886717 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-slash" (OuterVolumeSpecName: "host-slash") pod "04b20f02-6c1e-4082-8233-8f06bda63195" (UID: "04b20f02-6c1e-4082-8233-8f06bda63195"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886792 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/04b20f02-6c1e-4082-8233-8f06bda63195-ovn-node-metrics-cert\") pod \"04b20f02-6c1e-4082-8233-8f06bda63195\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886706 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "04b20f02-6c1e-4082-8233-8f06bda63195" (UID: "04b20f02-6c1e-4082-8233-8f06bda63195"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886814 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-systemd-units\") pod \"04b20f02-6c1e-4082-8233-8f06bda63195\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886836 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-kubelet\") pod \"04b20f02-6c1e-4082-8233-8f06bda63195\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886856 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-node-log\") pod \"04b20f02-6c1e-4082-8233-8f06bda63195\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886874 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/04b20f02-6c1e-4082-8233-8f06bda63195-ovnkube-config\") pod \"04b20f02-6c1e-4082-8233-8f06bda63195\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886889 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-run-openvswitch\") pod \"04b20f02-6c1e-4082-8233-8f06bda63195\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886761 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "04b20f02-6c1e-4082-8233-8f06bda63195" (UID: "04b20f02-6c1e-4082-8233-8f06bda63195"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886770 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "04b20f02-6c1e-4082-8233-8f06bda63195" (UID: "04b20f02-6c1e-4082-8233-8f06bda63195"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886853 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "04b20f02-6c1e-4082-8233-8f06bda63195" (UID: "04b20f02-6c1e-4082-8233-8f06bda63195"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886890 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "04b20f02-6c1e-4082-8233-8f06bda63195" (UID: "04b20f02-6c1e-4082-8233-8f06bda63195"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886957 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-log-socket" (OuterVolumeSpecName: "log-socket") pod "04b20f02-6c1e-4082-8233-8f06bda63195" (UID: "04b20f02-6c1e-4082-8233-8f06bda63195"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886908 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-log-socket\") pod \"04b20f02-6c1e-4082-8233-8f06bda63195\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.886981 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-node-log" (OuterVolumeSpecName: "node-log") pod "04b20f02-6c1e-4082-8233-8f06bda63195" (UID: "04b20f02-6c1e-4082-8233-8f06bda63195"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887020 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/04b20f02-6c1e-4082-8233-8f06bda63195-ovnkube-script-lib\") pod \"04b20f02-6c1e-4082-8233-8f06bda63195\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887060 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/04b20f02-6c1e-4082-8233-8f06bda63195-env-overrides\") pod \"04b20f02-6c1e-4082-8233-8f06bda63195\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887091 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-cni-bin\") pod \"04b20f02-6c1e-4082-8233-8f06bda63195\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887113 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-run-systemd\") pod \"04b20f02-6c1e-4082-8233-8f06bda63195\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887131 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-etc-openvswitch\") pod \"04b20f02-6c1e-4082-8233-8f06bda63195\" (UID: \"04b20f02-6c1e-4082-8233-8f06bda63195\") " Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887313 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a4febc2a-5af5-4acd-9521-527d275d2814-ovnkube-config\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887341 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04b20f02-6c1e-4082-8233-8f06bda63195-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "04b20f02-6c1e-4082-8233-8f06bda63195" (UID: "04b20f02-6c1e-4082-8233-8f06bda63195"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887353 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a4febc2a-5af5-4acd-9521-527d275d2814-ovn-node-metrics-cert\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887373 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "04b20f02-6c1e-4082-8233-8f06bda63195" (UID: "04b20f02-6c1e-4082-8233-8f06bda63195"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887377 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-run-ovn-kubernetes\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887399 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-var-lib-openvswitch\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887420 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-cni-netd\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887439 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-systemd-units\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887459 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a4febc2a-5af5-4acd-9521-527d275d2814-env-overrides\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887481 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-log-socket\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887502 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a4febc2a-5af5-4acd-9521-527d275d2814-ovnkube-script-lib\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887536 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-kubelet\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887571 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-run-ovn\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887614 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwd2j\" (UniqueName: \"kubernetes.io/projected/a4febc2a-5af5-4acd-9521-527d275d2814-kube-api-access-wwd2j\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887642 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-run-systemd\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887662 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-node-log\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887683 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-run-netns\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887680 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "04b20f02-6c1e-4082-8233-8f06bda63195" (UID: "04b20f02-6c1e-4082-8233-8f06bda63195"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887720 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-run-openvswitch\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887770 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-cni-bin\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887805 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887836 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-slash\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887854 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-etc-openvswitch\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887968 4714 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-systemd-units\") on node \"crc\" DevicePath \"\"" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.887987 4714 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-kubelet\") on node \"crc\" DevicePath \"\"" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.888003 4714 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/04b20f02-6c1e-4082-8233-8f06bda63195-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.888007 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "04b20f02-6c1e-4082-8233-8f06bda63195" (UID: "04b20f02-6c1e-4082-8233-8f06bda63195"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.888020 4714 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-node-log\") on node \"crc\" DevicePath \"\"" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.888099 4714 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-run-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.888131 4714 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-log-socket\") on node \"crc\" DevicePath \"\"" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.888149 4714 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.888245 4714 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.888265 4714 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.888281 4714 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-cni-netd\") on node \"crc\" DevicePath \"\"" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.888294 4714 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.888306 4714 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-slash\") on node \"crc\" DevicePath \"\"" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.888318 4714 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-run-netns\") on node \"crc\" DevicePath \"\"" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.888330 4714 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.888334 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04b20f02-6c1e-4082-8233-8f06bda63195-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "04b20f02-6c1e-4082-8233-8f06bda63195" (UID: "04b20f02-6c1e-4082-8233-8f06bda63195"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.888409 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04b20f02-6c1e-4082-8233-8f06bda63195-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "04b20f02-6c1e-4082-8233-8f06bda63195" (UID: "04b20f02-6c1e-4082-8233-8f06bda63195"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.892354 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04b20f02-6c1e-4082-8233-8f06bda63195-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "04b20f02-6c1e-4082-8233-8f06bda63195" (UID: "04b20f02-6c1e-4082-8233-8f06bda63195"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.892540 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04b20f02-6c1e-4082-8233-8f06bda63195-kube-api-access-7vrsm" (OuterVolumeSpecName: "kube-api-access-7vrsm") pod "04b20f02-6c1e-4082-8233-8f06bda63195" (UID: "04b20f02-6c1e-4082-8233-8f06bda63195"). InnerVolumeSpecName "kube-api-access-7vrsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.902983 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b2ttm_89560008-8bdc-4640-af11-681d825e69d4/kube-multus/2.log" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.903468 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b2ttm_89560008-8bdc-4640-af11-681d825e69d4/kube-multus/1.log" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.903508 4714 generic.go:334] "Generic (PLEG): container finished" podID="89560008-8bdc-4640-af11-681d825e69d4" containerID="e21aab3b653d9b1f38d58e9c32cbfb8988660ecb96eec4099a6536e09747d8fb" exitCode=2 Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.903557 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b2ttm" event={"ID":"89560008-8bdc-4640-af11-681d825e69d4","Type":"ContainerDied","Data":"e21aab3b653d9b1f38d58e9c32cbfb8988660ecb96eec4099a6536e09747d8fb"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.903591 4714 scope.go:117] "RemoveContainer" containerID="c747ed61a18e27d63630395860ce896242426b1ce46ea5f9d00534b808804a58" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.904059 4714 scope.go:117] "RemoveContainer" containerID="e21aab3b653d9b1f38d58e9c32cbfb8988660ecb96eec4099a6536e09747d8fb" Jan 29 16:21:49 crc kubenswrapper[4714]: E0129 16:21:49.904346 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-b2ttm_openshift-multus(89560008-8bdc-4640-af11-681d825e69d4)\"" pod="openshift-multus/multus-b2ttm" podUID="89560008-8bdc-4640-af11-681d825e69d4" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.905830 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-sbnkt_04b20f02-6c1e-4082-8233-8f06bda63195/ovnkube-controller/3.log" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.907365 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "04b20f02-6c1e-4082-8233-8f06bda63195" (UID: "04b20f02-6c1e-4082-8233-8f06bda63195"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.909293 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-sbnkt_04b20f02-6c1e-4082-8233-8f06bda63195/ovn-acl-logging/0.log" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.909734 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-sbnkt_04b20f02-6c1e-4082-8233-8f06bda63195/ovn-controller/0.log" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910104 4714 generic.go:334] "Generic (PLEG): container finished" podID="04b20f02-6c1e-4082-8233-8f06bda63195" containerID="00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6" exitCode=0 Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910176 4714 generic.go:334] "Generic (PLEG): container finished" podID="04b20f02-6c1e-4082-8233-8f06bda63195" containerID="662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf" exitCode=0 Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910175 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerDied","Data":"00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910195 4714 generic.go:334] "Generic (PLEG): container finished" podID="04b20f02-6c1e-4082-8233-8f06bda63195" containerID="5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7" exitCode=0 Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910208 4714 generic.go:334] "Generic (PLEG): container finished" podID="04b20f02-6c1e-4082-8233-8f06bda63195" containerID="e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4" exitCode=0 Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910217 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerDied","Data":"662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910235 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerDied","Data":"5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910247 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerDied","Data":"e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910258 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerDied","Data":"429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910219 4714 generic.go:334] "Generic (PLEG): container finished" podID="04b20f02-6c1e-4082-8233-8f06bda63195" containerID="429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816" exitCode=0 Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910271 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910277 4714 generic.go:334] "Generic (PLEG): container finished" podID="04b20f02-6c1e-4082-8233-8f06bda63195" containerID="b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc" exitCode=0 Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910366 4714 generic.go:334] "Generic (PLEG): container finished" podID="04b20f02-6c1e-4082-8233-8f06bda63195" containerID="2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd" exitCode=143 Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910375 4714 generic.go:334] "Generic (PLEG): container finished" podID="04b20f02-6c1e-4082-8233-8f06bda63195" containerID="7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016" exitCode=143 Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910286 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerDied","Data":"b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910408 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910422 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910431 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910438 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910446 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910453 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910460 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910467 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910474 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910481 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910499 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerDied","Data":"2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910513 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910522 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910529 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910536 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910543 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910551 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910558 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910566 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910573 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910580 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910590 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerDied","Data":"7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910601 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910609 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910616 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910623 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910630 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910637 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910644 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910651 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910659 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910665 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910677 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sbnkt" event={"ID":"04b20f02-6c1e-4082-8233-8f06bda63195","Type":"ContainerDied","Data":"17aad70fcdfcfc2aa07f37d1c4b0d894a800d6ca4c4b34e6100a73fad699fe31"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910695 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910714 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910725 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910733 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910740 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910747 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910754 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910761 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910768 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.910775 4714 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba"} Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.948622 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-sbnkt"] Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.951672 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-sbnkt"] Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.989629 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-kubelet\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.989692 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-run-ovn\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.989757 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwd2j\" (UniqueName: \"kubernetes.io/projected/a4febc2a-5af5-4acd-9521-527d275d2814-kube-api-access-wwd2j\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.989799 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-node-log\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.989793 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-kubelet\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.989819 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-run-systemd\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.989828 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-run-ovn\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.989848 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-run-netns\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.989900 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-run-netns\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.989906 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-run-systemd\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.989973 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-node-log\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990008 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-run-openvswitch\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990037 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-cni-bin\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990059 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990063 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-run-openvswitch\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990098 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-slash\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990113 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-cni-bin\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990120 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-etc-openvswitch\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990140 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-slash\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990097 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990153 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-etc-openvswitch\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990184 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a4febc2a-5af5-4acd-9521-527d275d2814-ovnkube-config\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990265 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a4febc2a-5af5-4acd-9521-527d275d2814-ovn-node-metrics-cert\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990302 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-run-ovn-kubernetes\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990327 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-var-lib-openvswitch\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990351 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-systemd-units\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990370 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-cni-netd\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990388 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a4febc2a-5af5-4acd-9521-527d275d2814-env-overrides\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990398 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-run-ovn-kubernetes\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990403 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-log-socket\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990442 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-var-lib-openvswitch\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990464 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-systemd-units\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990475 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a4febc2a-5af5-4acd-9521-527d275d2814-ovnkube-script-lib\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990490 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-host-cni-netd\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990427 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a4febc2a-5af5-4acd-9521-527d275d2814-log-socket\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990837 4714 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/04b20f02-6c1e-4082-8233-8f06bda63195-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990846 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a4febc2a-5af5-4acd-9521-527d275d2814-ovnkube-config\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990862 4714 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/04b20f02-6c1e-4082-8233-8f06bda63195-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990875 4714 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-host-cni-bin\") on node \"crc\" DevicePath \"\"" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990887 4714 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/04b20f02-6c1e-4082-8233-8f06bda63195-run-systemd\") on node \"crc\" DevicePath \"\"" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990902 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vrsm\" (UniqueName: \"kubernetes.io/projected/04b20f02-6c1e-4082-8233-8f06bda63195-kube-api-access-7vrsm\") on node \"crc\" DevicePath \"\"" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.990915 4714 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/04b20f02-6c1e-4082-8233-8f06bda63195-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.991251 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a4febc2a-5af5-4acd-9521-527d275d2814-env-overrides\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.991270 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a4febc2a-5af5-4acd-9521-527d275d2814-ovnkube-script-lib\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:49 crc kubenswrapper[4714]: I0129 16:21:49.994164 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a4febc2a-5af5-4acd-9521-527d275d2814-ovn-node-metrics-cert\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.010539 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwd2j\" (UniqueName: \"kubernetes.io/projected/a4febc2a-5af5-4acd-9521-527d275d2814-kube-api-access-wwd2j\") pod \"ovnkube-node-skzvq\" (UID: \"a4febc2a-5af5-4acd-9521-527d275d2814\") " pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.119102 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.199033 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04b20f02-6c1e-4082-8233-8f06bda63195" path="/var/lib/kubelet/pods/04b20f02-6c1e-4082-8233-8f06bda63195/volumes" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.211145 4714 scope.go:117] "RemoveContainer" containerID="00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.233391 4714 scope.go:117] "RemoveContainer" containerID="95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.249617 4714 scope.go:117] "RemoveContainer" containerID="662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.260608 4714 scope.go:117] "RemoveContainer" containerID="5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.274431 4714 scope.go:117] "RemoveContainer" containerID="e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.285249 4714 scope.go:117] "RemoveContainer" containerID="429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.302421 4714 scope.go:117] "RemoveContainer" containerID="b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.316396 4714 scope.go:117] "RemoveContainer" containerID="2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.334897 4714 scope.go:117] "RemoveContainer" containerID="7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.350519 4714 scope.go:117] "RemoveContainer" containerID="6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.422089 4714 scope.go:117] "RemoveContainer" containerID="00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6" Jan 29 16:21:50 crc kubenswrapper[4714]: E0129 16:21:50.422559 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6\": container with ID starting with 00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6 not found: ID does not exist" containerID="00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.422589 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6"} err="failed to get container status \"00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6\": rpc error: code = NotFound desc = could not find container \"00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6\": container with ID starting with 00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.422609 4714 scope.go:117] "RemoveContainer" containerID="95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365" Jan 29 16:21:50 crc kubenswrapper[4714]: E0129 16:21:50.423121 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365\": container with ID starting with 95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365 not found: ID does not exist" containerID="95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.423190 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365"} err="failed to get container status \"95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365\": rpc error: code = NotFound desc = could not find container \"95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365\": container with ID starting with 95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.423254 4714 scope.go:117] "RemoveContainer" containerID="662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf" Jan 29 16:21:50 crc kubenswrapper[4714]: E0129 16:21:50.423653 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\": container with ID starting with 662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf not found: ID does not exist" containerID="662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.423690 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf"} err="failed to get container status \"662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\": rpc error: code = NotFound desc = could not find container \"662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\": container with ID starting with 662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.423705 4714 scope.go:117] "RemoveContainer" containerID="5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7" Jan 29 16:21:50 crc kubenswrapper[4714]: E0129 16:21:50.424005 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\": container with ID starting with 5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7 not found: ID does not exist" containerID="5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.424039 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7"} err="failed to get container status \"5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\": rpc error: code = NotFound desc = could not find container \"5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\": container with ID starting with 5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.424101 4714 scope.go:117] "RemoveContainer" containerID="e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4" Jan 29 16:21:50 crc kubenswrapper[4714]: E0129 16:21:50.424353 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\": container with ID starting with e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4 not found: ID does not exist" containerID="e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.424379 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4"} err="failed to get container status \"e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\": rpc error: code = NotFound desc = could not find container \"e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\": container with ID starting with e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.424395 4714 scope.go:117] "RemoveContainer" containerID="429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816" Jan 29 16:21:50 crc kubenswrapper[4714]: E0129 16:21:50.424737 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\": container with ID starting with 429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816 not found: ID does not exist" containerID="429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.424767 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816"} err="failed to get container status \"429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\": rpc error: code = NotFound desc = could not find container \"429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\": container with ID starting with 429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.424785 4714 scope.go:117] "RemoveContainer" containerID="b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc" Jan 29 16:21:50 crc kubenswrapper[4714]: E0129 16:21:50.425055 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\": container with ID starting with b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc not found: ID does not exist" containerID="b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.425080 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc"} err="failed to get container status \"b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\": rpc error: code = NotFound desc = could not find container \"b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\": container with ID starting with b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.425093 4714 scope.go:117] "RemoveContainer" containerID="2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd" Jan 29 16:21:50 crc kubenswrapper[4714]: E0129 16:21:50.425346 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\": container with ID starting with 2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd not found: ID does not exist" containerID="2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.425379 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd"} err="failed to get container status \"2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\": rpc error: code = NotFound desc = could not find container \"2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\": container with ID starting with 2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.425432 4714 scope.go:117] "RemoveContainer" containerID="7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016" Jan 29 16:21:50 crc kubenswrapper[4714]: E0129 16:21:50.425652 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\": container with ID starting with 7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016 not found: ID does not exist" containerID="7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.425674 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016"} err="failed to get container status \"7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\": rpc error: code = NotFound desc = could not find container \"7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\": container with ID starting with 7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.425687 4714 scope.go:117] "RemoveContainer" containerID="6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba" Jan 29 16:21:50 crc kubenswrapper[4714]: E0129 16:21:50.425893 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\": container with ID starting with 6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba not found: ID does not exist" containerID="6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.425914 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba"} err="failed to get container status \"6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\": rpc error: code = NotFound desc = could not find container \"6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\": container with ID starting with 6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.425926 4714 scope.go:117] "RemoveContainer" containerID="00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.426158 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6"} err="failed to get container status \"00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6\": rpc error: code = NotFound desc = could not find container \"00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6\": container with ID starting with 00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.426183 4714 scope.go:117] "RemoveContainer" containerID="95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.426415 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365"} err="failed to get container status \"95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365\": rpc error: code = NotFound desc = could not find container \"95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365\": container with ID starting with 95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.426436 4714 scope.go:117] "RemoveContainer" containerID="662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.426611 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf"} err="failed to get container status \"662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\": rpc error: code = NotFound desc = could not find container \"662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\": container with ID starting with 662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.426629 4714 scope.go:117] "RemoveContainer" containerID="5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.426791 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7"} err="failed to get container status \"5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\": rpc error: code = NotFound desc = could not find container \"5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\": container with ID starting with 5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.426806 4714 scope.go:117] "RemoveContainer" containerID="e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.426975 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4"} err="failed to get container status \"e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\": rpc error: code = NotFound desc = could not find container \"e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\": container with ID starting with e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.426992 4714 scope.go:117] "RemoveContainer" containerID="429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.427171 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816"} err="failed to get container status \"429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\": rpc error: code = NotFound desc = could not find container \"429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\": container with ID starting with 429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.427187 4714 scope.go:117] "RemoveContainer" containerID="b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.428039 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc"} err="failed to get container status \"b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\": rpc error: code = NotFound desc = could not find container \"b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\": container with ID starting with b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.428059 4714 scope.go:117] "RemoveContainer" containerID="2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.428277 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd"} err="failed to get container status \"2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\": rpc error: code = NotFound desc = could not find container \"2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\": container with ID starting with 2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.428296 4714 scope.go:117] "RemoveContainer" containerID="7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.428456 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016"} err="failed to get container status \"7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\": rpc error: code = NotFound desc = could not find container \"7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\": container with ID starting with 7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.428474 4714 scope.go:117] "RemoveContainer" containerID="6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.428640 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba"} err="failed to get container status \"6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\": rpc error: code = NotFound desc = could not find container \"6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\": container with ID starting with 6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.428657 4714 scope.go:117] "RemoveContainer" containerID="00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.428810 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6"} err="failed to get container status \"00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6\": rpc error: code = NotFound desc = could not find container \"00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6\": container with ID starting with 00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.428826 4714 scope.go:117] "RemoveContainer" containerID="95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.429020 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365"} err="failed to get container status \"95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365\": rpc error: code = NotFound desc = could not find container \"95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365\": container with ID starting with 95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.429039 4714 scope.go:117] "RemoveContainer" containerID="662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.429373 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf"} err="failed to get container status \"662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\": rpc error: code = NotFound desc = could not find container \"662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\": container with ID starting with 662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.429390 4714 scope.go:117] "RemoveContainer" containerID="5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.429583 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7"} err="failed to get container status \"5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\": rpc error: code = NotFound desc = could not find container \"5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\": container with ID starting with 5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.429599 4714 scope.go:117] "RemoveContainer" containerID="e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.429829 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4"} err="failed to get container status \"e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\": rpc error: code = NotFound desc = could not find container \"e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\": container with ID starting with e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.429846 4714 scope.go:117] "RemoveContainer" containerID="429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.430029 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816"} err="failed to get container status \"429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\": rpc error: code = NotFound desc = could not find container \"429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\": container with ID starting with 429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.430046 4714 scope.go:117] "RemoveContainer" containerID="b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.430342 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc"} err="failed to get container status \"b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\": rpc error: code = NotFound desc = could not find container \"b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\": container with ID starting with b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.430368 4714 scope.go:117] "RemoveContainer" containerID="2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.431066 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd"} err="failed to get container status \"2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\": rpc error: code = NotFound desc = could not find container \"2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\": container with ID starting with 2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.431100 4714 scope.go:117] "RemoveContainer" containerID="7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.431483 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016"} err="failed to get container status \"7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\": rpc error: code = NotFound desc = could not find container \"7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\": container with ID starting with 7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.431535 4714 scope.go:117] "RemoveContainer" containerID="6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.431950 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba"} err="failed to get container status \"6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\": rpc error: code = NotFound desc = could not find container \"6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\": container with ID starting with 6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.432003 4714 scope.go:117] "RemoveContainer" containerID="00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.432382 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6"} err="failed to get container status \"00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6\": rpc error: code = NotFound desc = could not find container \"00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6\": container with ID starting with 00a2229f59557dd718e8e44cd8806fa686a96164db467c9fd584df1ca5f949c6 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.432408 4714 scope.go:117] "RemoveContainer" containerID="95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.432727 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365"} err="failed to get container status \"95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365\": rpc error: code = NotFound desc = could not find container \"95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365\": container with ID starting with 95506ff95d5b470923bf3c4615b45a7b4741260bb32a99839052875ecf50a365 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.432760 4714 scope.go:117] "RemoveContainer" containerID="662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.433086 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf"} err="failed to get container status \"662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\": rpc error: code = NotFound desc = could not find container \"662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf\": container with ID starting with 662852ce553d1fb58e5b7f129508aba9fa239b05fb7a0102edfe307908a49bdf not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.433112 4714 scope.go:117] "RemoveContainer" containerID="5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.433374 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7"} err="failed to get container status \"5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\": rpc error: code = NotFound desc = could not find container \"5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7\": container with ID starting with 5187dc8548f605295fbf37ce78f2d5a40a3ba0996c4d631fe6df1583db419df7 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.433411 4714 scope.go:117] "RemoveContainer" containerID="e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.434149 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4"} err="failed to get container status \"e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\": rpc error: code = NotFound desc = could not find container \"e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4\": container with ID starting with e625f30f0174bd89b2624c71963ed72e74cb0c7ad78f18998ec81ea57690fde4 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.434178 4714 scope.go:117] "RemoveContainer" containerID="429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.434434 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816"} err="failed to get container status \"429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\": rpc error: code = NotFound desc = could not find container \"429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816\": container with ID starting with 429e69f5ff2994413c83898a548cfdc9d11bd6a498838e0bc2fe8813f2800816 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.434462 4714 scope.go:117] "RemoveContainer" containerID="b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.434703 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc"} err="failed to get container status \"b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\": rpc error: code = NotFound desc = could not find container \"b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc\": container with ID starting with b1d444ab76cebea1c9d94e3caae9a82ce1a3f9b5b98eed31548472a8f8f6c4bc not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.434749 4714 scope.go:117] "RemoveContainer" containerID="2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.435052 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd"} err="failed to get container status \"2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\": rpc error: code = NotFound desc = could not find container \"2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd\": container with ID starting with 2a5f6317ab204866e3d4a05f50396b0b2786b0e1d922d255c3a2b53c7e6968fd not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.435080 4714 scope.go:117] "RemoveContainer" containerID="7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.435324 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016"} err="failed to get container status \"7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\": rpc error: code = NotFound desc = could not find container \"7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016\": container with ID starting with 7d00f0f27a2d4d8a5e65608561bbad3e2d2017afa341e78c2914e2061f983016 not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.435348 4714 scope.go:117] "RemoveContainer" containerID="6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.435582 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba"} err="failed to get container status \"6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\": rpc error: code = NotFound desc = could not find container \"6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba\": container with ID starting with 6025ce34465bd94b02e72f82b958993a2aa8c0ed695f7b6c70f8dd4475ec67ba not found: ID does not exist" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.916920 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b2ttm_89560008-8bdc-4640-af11-681d825e69d4/kube-multus/2.log" Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.919179 4714 generic.go:334] "Generic (PLEG): container finished" podID="a4febc2a-5af5-4acd-9521-527d275d2814" containerID="041701549e9197314050e0918afcd67cc692ec700b123e479cc8b7d528218f5a" exitCode=0 Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.919281 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" event={"ID":"a4febc2a-5af5-4acd-9521-527d275d2814","Type":"ContainerDied","Data":"041701549e9197314050e0918afcd67cc692ec700b123e479cc8b7d528218f5a"} Jan 29 16:21:50 crc kubenswrapper[4714]: I0129 16:21:50.919334 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" event={"ID":"a4febc2a-5af5-4acd-9521-527d275d2814","Type":"ContainerStarted","Data":"a4e87bf80e9415e77e846fb0756debf13b4e453e2713dec1043b3d31093ccf48"} Jan 29 16:21:51 crc kubenswrapper[4714]: I0129 16:21:51.932250 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" event={"ID":"a4febc2a-5af5-4acd-9521-527d275d2814","Type":"ContainerStarted","Data":"b1478134ba37f7e9fe2f8f1c3a7614a4176990c9d63d024ea6a6366a9ffd5ef4"} Jan 29 16:21:51 crc kubenswrapper[4714]: I0129 16:21:51.932685 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" event={"ID":"a4febc2a-5af5-4acd-9521-527d275d2814","Type":"ContainerStarted","Data":"cf40f584012c0ab702c5121a367f19e69f14be7138e56a223b208622c970c7fe"} Jan 29 16:21:51 crc kubenswrapper[4714]: I0129 16:21:51.932711 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" event={"ID":"a4febc2a-5af5-4acd-9521-527d275d2814","Type":"ContainerStarted","Data":"8f203e1c5feaedd818df2c932055b9ca8e04a4e3d2a22c8101529b4b66c37e5e"} Jan 29 16:21:52 crc kubenswrapper[4714]: I0129 16:21:52.942435 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" event={"ID":"a4febc2a-5af5-4acd-9521-527d275d2814","Type":"ContainerStarted","Data":"5e8ee4e72c5e7be52b0a982526799ee625e6153e19584ad2162494477e76aceb"} Jan 29 16:21:52 crc kubenswrapper[4714]: I0129 16:21:52.942500 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" event={"ID":"a4febc2a-5af5-4acd-9521-527d275d2814","Type":"ContainerStarted","Data":"81152bf9ad2c03871e447d73002e05950a372535404748d0a5554458e8dbbc0a"} Jan 29 16:21:52 crc kubenswrapper[4714]: I0129 16:21:52.942518 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" event={"ID":"a4febc2a-5af5-4acd-9521-527d275d2814","Type":"ContainerStarted","Data":"d3c6c9a6cbe7093656147cf42544feb7f3f26a5e115340d9fb9c9d94206090ae"} Jan 29 16:21:54 crc kubenswrapper[4714]: I0129 16:21:54.960213 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" event={"ID":"a4febc2a-5af5-4acd-9521-527d275d2814","Type":"ContainerStarted","Data":"aa8ac0cdf264663cc0ef0807cffc7f26644661fa11d01105a502295dcd648ca6"} Jan 29 16:21:56 crc kubenswrapper[4714]: I0129 16:21:56.974368 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" event={"ID":"a4febc2a-5af5-4acd-9521-527d275d2814","Type":"ContainerStarted","Data":"01a66910506befdef6b42002992c445aa79f1addfe47a86c32f58e266f3df838"} Jan 29 16:21:56 crc kubenswrapper[4714]: I0129 16:21:56.975008 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:56 crc kubenswrapper[4714]: I0129 16:21:56.975025 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:57 crc kubenswrapper[4714]: I0129 16:21:57.005804 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:57 crc kubenswrapper[4714]: I0129 16:21:57.011029 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" podStartSLOduration=8.011013825 podStartE2EDuration="8.011013825s" podCreationTimestamp="2026-01-29 16:21:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:21:57.007482218 +0000 UTC m=+723.527983338" watchObservedRunningTime="2026-01-29 16:21:57.011013825 +0000 UTC m=+723.531514935" Jan 29 16:21:57 crc kubenswrapper[4714]: I0129 16:21:57.981868 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:21:58 crc kubenswrapper[4714]: I0129 16:21:58.014155 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:22:03 crc kubenswrapper[4714]: I0129 16:22:03.184106 4714 scope.go:117] "RemoveContainer" containerID="e21aab3b653d9b1f38d58e9c32cbfb8988660ecb96eec4099a6536e09747d8fb" Jan 29 16:22:03 crc kubenswrapper[4714]: E0129 16:22:03.184889 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-b2ttm_openshift-multus(89560008-8bdc-4640-af11-681d825e69d4)\"" pod="openshift-multus/multus-b2ttm" podUID="89560008-8bdc-4640-af11-681d825e69d4" Jan 29 16:22:12 crc kubenswrapper[4714]: I0129 16:22:12.001551 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st"] Jan 29 16:22:12 crc kubenswrapper[4714]: I0129 16:22:12.004106 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:12 crc kubenswrapper[4714]: I0129 16:22:12.006827 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 29 16:22:12 crc kubenswrapper[4714]: I0129 16:22:12.024323 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st"] Jan 29 16:22:12 crc kubenswrapper[4714]: I0129 16:22:12.087789 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c184c6f2-1af5-4f70-9251-6beb2baae06b-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st\" (UID: \"c184c6f2-1af5-4f70-9251-6beb2baae06b\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:12 crc kubenswrapper[4714]: I0129 16:22:12.087873 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjp7x\" (UniqueName: \"kubernetes.io/projected/c184c6f2-1af5-4f70-9251-6beb2baae06b-kube-api-access-zjp7x\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st\" (UID: \"c184c6f2-1af5-4f70-9251-6beb2baae06b\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:12 crc kubenswrapper[4714]: I0129 16:22:12.088044 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c184c6f2-1af5-4f70-9251-6beb2baae06b-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st\" (UID: \"c184c6f2-1af5-4f70-9251-6beb2baae06b\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:12 crc kubenswrapper[4714]: I0129 16:22:12.188874 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c184c6f2-1af5-4f70-9251-6beb2baae06b-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st\" (UID: \"c184c6f2-1af5-4f70-9251-6beb2baae06b\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:12 crc kubenswrapper[4714]: I0129 16:22:12.189111 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c184c6f2-1af5-4f70-9251-6beb2baae06b-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st\" (UID: \"c184c6f2-1af5-4f70-9251-6beb2baae06b\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:12 crc kubenswrapper[4714]: I0129 16:22:12.189193 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjp7x\" (UniqueName: \"kubernetes.io/projected/c184c6f2-1af5-4f70-9251-6beb2baae06b-kube-api-access-zjp7x\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st\" (UID: \"c184c6f2-1af5-4f70-9251-6beb2baae06b\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:12 crc kubenswrapper[4714]: I0129 16:22:12.189859 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c184c6f2-1af5-4f70-9251-6beb2baae06b-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st\" (UID: \"c184c6f2-1af5-4f70-9251-6beb2baae06b\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:12 crc kubenswrapper[4714]: I0129 16:22:12.189994 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c184c6f2-1af5-4f70-9251-6beb2baae06b-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st\" (UID: \"c184c6f2-1af5-4f70-9251-6beb2baae06b\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:12 crc kubenswrapper[4714]: I0129 16:22:12.224993 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjp7x\" (UniqueName: \"kubernetes.io/projected/c184c6f2-1af5-4f70-9251-6beb2baae06b-kube-api-access-zjp7x\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st\" (UID: \"c184c6f2-1af5-4f70-9251-6beb2baae06b\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:12 crc kubenswrapper[4714]: I0129 16:22:12.338206 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:12 crc kubenswrapper[4714]: E0129 16:22:12.374710 4714 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_openshift-marketplace_c184c6f2-1af5-4f70-9251-6beb2baae06b_0(5956f26abe33ed8221a24b1daa95878a91ce67788da634e2a0ca4f98e920f958): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 29 16:22:12 crc kubenswrapper[4714]: E0129 16:22:12.374802 4714 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_openshift-marketplace_c184c6f2-1af5-4f70-9251-6beb2baae06b_0(5956f26abe33ed8221a24b1daa95878a91ce67788da634e2a0ca4f98e920f958): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:12 crc kubenswrapper[4714]: E0129 16:22:12.374829 4714 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_openshift-marketplace_c184c6f2-1af5-4f70-9251-6beb2baae06b_0(5956f26abe33ed8221a24b1daa95878a91ce67788da634e2a0ca4f98e920f958): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:12 crc kubenswrapper[4714]: E0129 16:22:12.374893 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_openshift-marketplace(c184c6f2-1af5-4f70-9251-6beb2baae06b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_openshift-marketplace(c184c6f2-1af5-4f70-9251-6beb2baae06b)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_openshift-marketplace_c184c6f2-1af5-4f70-9251-6beb2baae06b_0(5956f26abe33ed8221a24b1daa95878a91ce67788da634e2a0ca4f98e920f958): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" podUID="c184c6f2-1af5-4f70-9251-6beb2baae06b" Jan 29 16:22:13 crc kubenswrapper[4714]: I0129 16:22:13.092248 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:13 crc kubenswrapper[4714]: I0129 16:22:13.092794 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:13 crc kubenswrapper[4714]: E0129 16:22:13.112739 4714 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_openshift-marketplace_c184c6f2-1af5-4f70-9251-6beb2baae06b_0(30e6d6bea70f12f3450627a5bc192b45822df142fc1968ef1a5441b4b22377b0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 29 16:22:13 crc kubenswrapper[4714]: E0129 16:22:13.112848 4714 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_openshift-marketplace_c184c6f2-1af5-4f70-9251-6beb2baae06b_0(30e6d6bea70f12f3450627a5bc192b45822df142fc1968ef1a5441b4b22377b0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:13 crc kubenswrapper[4714]: E0129 16:22:13.112888 4714 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_openshift-marketplace_c184c6f2-1af5-4f70-9251-6beb2baae06b_0(30e6d6bea70f12f3450627a5bc192b45822df142fc1968ef1a5441b4b22377b0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:13 crc kubenswrapper[4714]: E0129 16:22:13.112974 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_openshift-marketplace(c184c6f2-1af5-4f70-9251-6beb2baae06b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_openshift-marketplace(c184c6f2-1af5-4f70-9251-6beb2baae06b)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_openshift-marketplace_c184c6f2-1af5-4f70-9251-6beb2baae06b_0(30e6d6bea70f12f3450627a5bc192b45822df142fc1968ef1a5441b4b22377b0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" podUID="c184c6f2-1af5-4f70-9251-6beb2baae06b" Jan 29 16:22:15 crc kubenswrapper[4714]: I0129 16:22:15.185153 4714 scope.go:117] "RemoveContainer" containerID="e21aab3b653d9b1f38d58e9c32cbfb8988660ecb96eec4099a6536e09747d8fb" Jan 29 16:22:16 crc kubenswrapper[4714]: I0129 16:22:16.136726 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b2ttm_89560008-8bdc-4640-af11-681d825e69d4/kube-multus/2.log" Jan 29 16:22:16 crc kubenswrapper[4714]: I0129 16:22:16.137327 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b2ttm" event={"ID":"89560008-8bdc-4640-af11-681d825e69d4","Type":"ContainerStarted","Data":"4caf97375d8de0e64eae8f7542cedd216d5913d276a7c525626882159c3c130b"} Jan 29 16:22:20 crc kubenswrapper[4714]: I0129 16:22:20.149974 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-skzvq" Jan 29 16:22:23 crc kubenswrapper[4714]: I0129 16:22:23.183611 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:23 crc kubenswrapper[4714]: I0129 16:22:23.185147 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:23 crc kubenswrapper[4714]: I0129 16:22:23.389978 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st"] Jan 29 16:22:24 crc kubenswrapper[4714]: I0129 16:22:24.190894 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" event={"ID":"c184c6f2-1af5-4f70-9251-6beb2baae06b","Type":"ContainerStarted","Data":"869accf0303863a05a34ac3c3845988fadd47bff208f4ed12415981d69ec342a"} Jan 29 16:22:26 crc kubenswrapper[4714]: I0129 16:22:26.203923 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" event={"ID":"c184c6f2-1af5-4f70-9251-6beb2baae06b","Type":"ContainerStarted","Data":"05e0a9023a315c8e823d2a29c8e0d2e37a13cd8fbf2c97a5e88c1825840657f8"} Jan 29 16:22:27 crc kubenswrapper[4714]: I0129 16:22:27.211661 4714 generic.go:334] "Generic (PLEG): container finished" podID="c184c6f2-1af5-4f70-9251-6beb2baae06b" containerID="05e0a9023a315c8e823d2a29c8e0d2e37a13cd8fbf2c97a5e88c1825840657f8" exitCode=0 Jan 29 16:22:27 crc kubenswrapper[4714]: I0129 16:22:27.211821 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" event={"ID":"c184c6f2-1af5-4f70-9251-6beb2baae06b","Type":"ContainerDied","Data":"05e0a9023a315c8e823d2a29c8e0d2e37a13cd8fbf2c97a5e88c1825840657f8"} Jan 29 16:22:27 crc kubenswrapper[4714]: I0129 16:22:27.213985 4714 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 16:22:32 crc kubenswrapper[4714]: I0129 16:22:32.239871 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" event={"ID":"c184c6f2-1af5-4f70-9251-6beb2baae06b","Type":"ContainerStarted","Data":"fb8c01e794f29fc68495ebe7c0293de3d5b3be2b85ddbfd0e228c600da4a1d06"} Jan 29 16:22:32 crc kubenswrapper[4714]: I0129 16:22:32.821700 4714 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 29 16:22:33 crc kubenswrapper[4714]: I0129 16:22:33.246440 4714 generic.go:334] "Generic (PLEG): container finished" podID="c184c6f2-1af5-4f70-9251-6beb2baae06b" containerID="fb8c01e794f29fc68495ebe7c0293de3d5b3be2b85ddbfd0e228c600da4a1d06" exitCode=0 Jan 29 16:22:33 crc kubenswrapper[4714]: I0129 16:22:33.246496 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" event={"ID":"c184c6f2-1af5-4f70-9251-6beb2baae06b","Type":"ContainerDied","Data":"fb8c01e794f29fc68495ebe7c0293de3d5b3be2b85ddbfd0e228c600da4a1d06"} Jan 29 16:22:33 crc kubenswrapper[4714]: I0129 16:22:33.805243 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9ttgj"] Jan 29 16:22:33 crc kubenswrapper[4714]: I0129 16:22:33.807505 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9ttgj" Jan 29 16:22:33 crc kubenswrapper[4714]: I0129 16:22:33.826549 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9ttgj"] Jan 29 16:22:33 crc kubenswrapper[4714]: I0129 16:22:33.902497 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b93b136-c182-4b20-89a4-1d61e1d2d03c-utilities\") pod \"redhat-operators-9ttgj\" (UID: \"3b93b136-c182-4b20-89a4-1d61e1d2d03c\") " pod="openshift-marketplace/redhat-operators-9ttgj" Jan 29 16:22:33 crc kubenswrapper[4714]: I0129 16:22:33.902651 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkw5r\" (UniqueName: \"kubernetes.io/projected/3b93b136-c182-4b20-89a4-1d61e1d2d03c-kube-api-access-xkw5r\") pod \"redhat-operators-9ttgj\" (UID: \"3b93b136-c182-4b20-89a4-1d61e1d2d03c\") " pod="openshift-marketplace/redhat-operators-9ttgj" Jan 29 16:22:33 crc kubenswrapper[4714]: I0129 16:22:33.902719 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b93b136-c182-4b20-89a4-1d61e1d2d03c-catalog-content\") pod \"redhat-operators-9ttgj\" (UID: \"3b93b136-c182-4b20-89a4-1d61e1d2d03c\") " pod="openshift-marketplace/redhat-operators-9ttgj" Jan 29 16:22:34 crc kubenswrapper[4714]: I0129 16:22:34.004209 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b93b136-c182-4b20-89a4-1d61e1d2d03c-utilities\") pod \"redhat-operators-9ttgj\" (UID: \"3b93b136-c182-4b20-89a4-1d61e1d2d03c\") " pod="openshift-marketplace/redhat-operators-9ttgj" Jan 29 16:22:34 crc kubenswrapper[4714]: I0129 16:22:34.004380 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkw5r\" (UniqueName: \"kubernetes.io/projected/3b93b136-c182-4b20-89a4-1d61e1d2d03c-kube-api-access-xkw5r\") pod \"redhat-operators-9ttgj\" (UID: \"3b93b136-c182-4b20-89a4-1d61e1d2d03c\") " pod="openshift-marketplace/redhat-operators-9ttgj" Jan 29 16:22:34 crc kubenswrapper[4714]: I0129 16:22:34.004443 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b93b136-c182-4b20-89a4-1d61e1d2d03c-catalog-content\") pod \"redhat-operators-9ttgj\" (UID: \"3b93b136-c182-4b20-89a4-1d61e1d2d03c\") " pod="openshift-marketplace/redhat-operators-9ttgj" Jan 29 16:22:34 crc kubenswrapper[4714]: I0129 16:22:34.004900 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b93b136-c182-4b20-89a4-1d61e1d2d03c-utilities\") pod \"redhat-operators-9ttgj\" (UID: \"3b93b136-c182-4b20-89a4-1d61e1d2d03c\") " pod="openshift-marketplace/redhat-operators-9ttgj" Jan 29 16:22:34 crc kubenswrapper[4714]: I0129 16:22:34.005072 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b93b136-c182-4b20-89a4-1d61e1d2d03c-catalog-content\") pod \"redhat-operators-9ttgj\" (UID: \"3b93b136-c182-4b20-89a4-1d61e1d2d03c\") " pod="openshift-marketplace/redhat-operators-9ttgj" Jan 29 16:22:34 crc kubenswrapper[4714]: I0129 16:22:34.021899 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkw5r\" (UniqueName: \"kubernetes.io/projected/3b93b136-c182-4b20-89a4-1d61e1d2d03c-kube-api-access-xkw5r\") pod \"redhat-operators-9ttgj\" (UID: \"3b93b136-c182-4b20-89a4-1d61e1d2d03c\") " pod="openshift-marketplace/redhat-operators-9ttgj" Jan 29 16:22:34 crc kubenswrapper[4714]: I0129 16:22:34.145221 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9ttgj" Jan 29 16:22:34 crc kubenswrapper[4714]: I0129 16:22:34.254025 4714 generic.go:334] "Generic (PLEG): container finished" podID="c184c6f2-1af5-4f70-9251-6beb2baae06b" containerID="c3c80ba39f2166deea58a208aff618c1c394a583efafcb8a7f94ec8b247ab86e" exitCode=0 Jan 29 16:22:34 crc kubenswrapper[4714]: I0129 16:22:34.254081 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" event={"ID":"c184c6f2-1af5-4f70-9251-6beb2baae06b","Type":"ContainerDied","Data":"c3c80ba39f2166deea58a208aff618c1c394a583efafcb8a7f94ec8b247ab86e"} Jan 29 16:22:34 crc kubenswrapper[4714]: I0129 16:22:34.430840 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9ttgj"] Jan 29 16:22:35 crc kubenswrapper[4714]: I0129 16:22:35.260958 4714 generic.go:334] "Generic (PLEG): container finished" podID="3b93b136-c182-4b20-89a4-1d61e1d2d03c" containerID="dc5891a8c29afb2826ce230ab3fd28a92cd17c413c6d5d58f4c0d944c000e9ef" exitCode=0 Jan 29 16:22:35 crc kubenswrapper[4714]: I0129 16:22:35.261031 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ttgj" event={"ID":"3b93b136-c182-4b20-89a4-1d61e1d2d03c","Type":"ContainerDied","Data":"dc5891a8c29afb2826ce230ab3fd28a92cd17c413c6d5d58f4c0d944c000e9ef"} Jan 29 16:22:35 crc kubenswrapper[4714]: I0129 16:22:35.261319 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ttgj" event={"ID":"3b93b136-c182-4b20-89a4-1d61e1d2d03c","Type":"ContainerStarted","Data":"4716263f07ca6d3ad4b5be1c8fb4857b1c649402f03ef208b20007f3c151a261"} Jan 29 16:22:35 crc kubenswrapper[4714]: I0129 16:22:35.550120 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:35 crc kubenswrapper[4714]: I0129 16:22:35.724979 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjp7x\" (UniqueName: \"kubernetes.io/projected/c184c6f2-1af5-4f70-9251-6beb2baae06b-kube-api-access-zjp7x\") pod \"c184c6f2-1af5-4f70-9251-6beb2baae06b\" (UID: \"c184c6f2-1af5-4f70-9251-6beb2baae06b\") " Jan 29 16:22:35 crc kubenswrapper[4714]: I0129 16:22:35.725059 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c184c6f2-1af5-4f70-9251-6beb2baae06b-bundle\") pod \"c184c6f2-1af5-4f70-9251-6beb2baae06b\" (UID: \"c184c6f2-1af5-4f70-9251-6beb2baae06b\") " Jan 29 16:22:35 crc kubenswrapper[4714]: I0129 16:22:35.725150 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c184c6f2-1af5-4f70-9251-6beb2baae06b-util\") pod \"c184c6f2-1af5-4f70-9251-6beb2baae06b\" (UID: \"c184c6f2-1af5-4f70-9251-6beb2baae06b\") " Jan 29 16:22:35 crc kubenswrapper[4714]: I0129 16:22:35.726128 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c184c6f2-1af5-4f70-9251-6beb2baae06b-bundle" (OuterVolumeSpecName: "bundle") pod "c184c6f2-1af5-4f70-9251-6beb2baae06b" (UID: "c184c6f2-1af5-4f70-9251-6beb2baae06b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:22:35 crc kubenswrapper[4714]: I0129 16:22:35.734103 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c184c6f2-1af5-4f70-9251-6beb2baae06b-kube-api-access-zjp7x" (OuterVolumeSpecName: "kube-api-access-zjp7x") pod "c184c6f2-1af5-4f70-9251-6beb2baae06b" (UID: "c184c6f2-1af5-4f70-9251-6beb2baae06b"). InnerVolumeSpecName "kube-api-access-zjp7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:22:35 crc kubenswrapper[4714]: I0129 16:22:35.738948 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c184c6f2-1af5-4f70-9251-6beb2baae06b-util" (OuterVolumeSpecName: "util") pod "c184c6f2-1af5-4f70-9251-6beb2baae06b" (UID: "c184c6f2-1af5-4f70-9251-6beb2baae06b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:22:35 crc kubenswrapper[4714]: I0129 16:22:35.827189 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjp7x\" (UniqueName: \"kubernetes.io/projected/c184c6f2-1af5-4f70-9251-6beb2baae06b-kube-api-access-zjp7x\") on node \"crc\" DevicePath \"\"" Jan 29 16:22:35 crc kubenswrapper[4714]: I0129 16:22:35.827249 4714 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c184c6f2-1af5-4f70-9251-6beb2baae06b-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 16:22:35 crc kubenswrapper[4714]: I0129 16:22:35.827274 4714 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c184c6f2-1af5-4f70-9251-6beb2baae06b-util\") on node \"crc\" DevicePath \"\"" Jan 29 16:22:36 crc kubenswrapper[4714]: I0129 16:22:36.279873 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" event={"ID":"c184c6f2-1af5-4f70-9251-6beb2baae06b","Type":"ContainerDied","Data":"869accf0303863a05a34ac3c3845988fadd47bff208f4ed12415981d69ec342a"} Jan 29 16:22:36 crc kubenswrapper[4714]: I0129 16:22:36.279957 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="869accf0303863a05a34ac3c3845988fadd47bff208f4ed12415981d69ec342a" Jan 29 16:22:36 crc kubenswrapper[4714]: I0129 16:22:36.280186 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st" Jan 29 16:22:38 crc kubenswrapper[4714]: I0129 16:22:38.294188 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ttgj" event={"ID":"3b93b136-c182-4b20-89a4-1d61e1d2d03c","Type":"ContainerStarted","Data":"6632448df0e09a8f0e5148074bcd76d6b08015e4e356eeae9a581a934dfb3cac"} Jan 29 16:22:39 crc kubenswrapper[4714]: I0129 16:22:39.302108 4714 generic.go:334] "Generic (PLEG): container finished" podID="3b93b136-c182-4b20-89a4-1d61e1d2d03c" containerID="6632448df0e09a8f0e5148074bcd76d6b08015e4e356eeae9a581a934dfb3cac" exitCode=0 Jan 29 16:22:39 crc kubenswrapper[4714]: I0129 16:22:39.302225 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ttgj" event={"ID":"3b93b136-c182-4b20-89a4-1d61e1d2d03c","Type":"ContainerDied","Data":"6632448df0e09a8f0e5148074bcd76d6b08015e4e356eeae9a581a934dfb3cac"} Jan 29 16:22:42 crc kubenswrapper[4714]: I0129 16:22:42.319363 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ttgj" event={"ID":"3b93b136-c182-4b20-89a4-1d61e1d2d03c","Type":"ContainerStarted","Data":"71891b59c66ba86ccba3be13ccefa3eee4b9a927ebc16ecd2fd3f3077bf4cab2"} Jan 29 16:22:42 crc kubenswrapper[4714]: I0129 16:22:42.340139 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9ttgj" podStartSLOduration=3.49502799 podStartE2EDuration="9.340121388s" podCreationTimestamp="2026-01-29 16:22:33 +0000 UTC" firstStartedPulling="2026-01-29 16:22:35.26318497 +0000 UTC m=+761.783686090" lastFinishedPulling="2026-01-29 16:22:41.108278348 +0000 UTC m=+767.628779488" observedRunningTime="2026-01-29 16:22:42.336464397 +0000 UTC m=+768.856965517" watchObservedRunningTime="2026-01-29 16:22:42.340121388 +0000 UTC m=+768.860622508" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.543991 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-586b87b897-zpr4q"] Jan 29 16:22:43 crc kubenswrapper[4714]: E0129 16:22:43.544197 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c184c6f2-1af5-4f70-9251-6beb2baae06b" containerName="extract" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.544209 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="c184c6f2-1af5-4f70-9251-6beb2baae06b" containerName="extract" Jan 29 16:22:43 crc kubenswrapper[4714]: E0129 16:22:43.544218 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c184c6f2-1af5-4f70-9251-6beb2baae06b" containerName="util" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.544224 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="c184c6f2-1af5-4f70-9251-6beb2baae06b" containerName="util" Jan 29 16:22:43 crc kubenswrapper[4714]: E0129 16:22:43.544242 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c184c6f2-1af5-4f70-9251-6beb2baae06b" containerName="pull" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.544247 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="c184c6f2-1af5-4f70-9251-6beb2baae06b" containerName="pull" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.544333 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="c184c6f2-1af5-4f70-9251-6beb2baae06b" containerName="extract" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.544712 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-586b87b897-zpr4q" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.547818 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.548632 4714 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-d6tsl" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.548781 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.548880 4714 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.549666 4714 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.568481 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-586b87b897-zpr4q"] Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.720026 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/432a4f98-877c-4f7a-b2b0-ce273a77450a-apiservice-cert\") pod \"metallb-operator-controller-manager-586b87b897-zpr4q\" (UID: \"432a4f98-877c-4f7a-b2b0-ce273a77450a\") " pod="metallb-system/metallb-operator-controller-manager-586b87b897-zpr4q" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.720402 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/432a4f98-877c-4f7a-b2b0-ce273a77450a-webhook-cert\") pod \"metallb-operator-controller-manager-586b87b897-zpr4q\" (UID: \"432a4f98-877c-4f7a-b2b0-ce273a77450a\") " pod="metallb-system/metallb-operator-controller-manager-586b87b897-zpr4q" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.720444 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9llmx\" (UniqueName: \"kubernetes.io/projected/432a4f98-877c-4f7a-b2b0-ce273a77450a-kube-api-access-9llmx\") pod \"metallb-operator-controller-manager-586b87b897-zpr4q\" (UID: \"432a4f98-877c-4f7a-b2b0-ce273a77450a\") " pod="metallb-system/metallb-operator-controller-manager-586b87b897-zpr4q" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.777828 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7df7c8d444-xs67n"] Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.778516 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7df7c8d444-xs67n" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.781291 4714 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-r42vn" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.781314 4714 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.781451 4714 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.821353 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/432a4f98-877c-4f7a-b2b0-ce273a77450a-apiservice-cert\") pod \"metallb-operator-controller-manager-586b87b897-zpr4q\" (UID: \"432a4f98-877c-4f7a-b2b0-ce273a77450a\") " pod="metallb-system/metallb-operator-controller-manager-586b87b897-zpr4q" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.821634 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/432a4f98-877c-4f7a-b2b0-ce273a77450a-webhook-cert\") pod \"metallb-operator-controller-manager-586b87b897-zpr4q\" (UID: \"432a4f98-877c-4f7a-b2b0-ce273a77450a\") " pod="metallb-system/metallb-operator-controller-manager-586b87b897-zpr4q" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.821731 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9llmx\" (UniqueName: \"kubernetes.io/projected/432a4f98-877c-4f7a-b2b0-ce273a77450a-kube-api-access-9llmx\") pod \"metallb-operator-controller-manager-586b87b897-zpr4q\" (UID: \"432a4f98-877c-4f7a-b2b0-ce273a77450a\") " pod="metallb-system/metallb-operator-controller-manager-586b87b897-zpr4q" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.826735 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/432a4f98-877c-4f7a-b2b0-ce273a77450a-apiservice-cert\") pod \"metallb-operator-controller-manager-586b87b897-zpr4q\" (UID: \"432a4f98-877c-4f7a-b2b0-ce273a77450a\") " pod="metallb-system/metallb-operator-controller-manager-586b87b897-zpr4q" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.830799 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/432a4f98-877c-4f7a-b2b0-ce273a77450a-webhook-cert\") pod \"metallb-operator-controller-manager-586b87b897-zpr4q\" (UID: \"432a4f98-877c-4f7a-b2b0-ce273a77450a\") " pod="metallb-system/metallb-operator-controller-manager-586b87b897-zpr4q" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.837673 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7df7c8d444-xs67n"] Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.849534 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9llmx\" (UniqueName: \"kubernetes.io/projected/432a4f98-877c-4f7a-b2b0-ce273a77450a-kube-api-access-9llmx\") pod \"metallb-operator-controller-manager-586b87b897-zpr4q\" (UID: \"432a4f98-877c-4f7a-b2b0-ce273a77450a\") " pod="metallb-system/metallb-operator-controller-manager-586b87b897-zpr4q" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.859831 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-586b87b897-zpr4q" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.923190 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ffe179b8-a1c8-430b-94f5-920aacf0defe-webhook-cert\") pod \"metallb-operator-webhook-server-7df7c8d444-xs67n\" (UID: \"ffe179b8-a1c8-430b-94f5-920aacf0defe\") " pod="metallb-system/metallb-operator-webhook-server-7df7c8d444-xs67n" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.923327 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q99dn\" (UniqueName: \"kubernetes.io/projected/ffe179b8-a1c8-430b-94f5-920aacf0defe-kube-api-access-q99dn\") pod \"metallb-operator-webhook-server-7df7c8d444-xs67n\" (UID: \"ffe179b8-a1c8-430b-94f5-920aacf0defe\") " pod="metallb-system/metallb-operator-webhook-server-7df7c8d444-xs67n" Jan 29 16:22:43 crc kubenswrapper[4714]: I0129 16:22:43.923383 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ffe179b8-a1c8-430b-94f5-920aacf0defe-apiservice-cert\") pod \"metallb-operator-webhook-server-7df7c8d444-xs67n\" (UID: \"ffe179b8-a1c8-430b-94f5-920aacf0defe\") " pod="metallb-system/metallb-operator-webhook-server-7df7c8d444-xs67n" Jan 29 16:22:44 crc kubenswrapper[4714]: I0129 16:22:44.032468 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ffe179b8-a1c8-430b-94f5-920aacf0defe-apiservice-cert\") pod \"metallb-operator-webhook-server-7df7c8d444-xs67n\" (UID: \"ffe179b8-a1c8-430b-94f5-920aacf0defe\") " pod="metallb-system/metallb-operator-webhook-server-7df7c8d444-xs67n" Jan 29 16:22:44 crc kubenswrapper[4714]: I0129 16:22:44.032828 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ffe179b8-a1c8-430b-94f5-920aacf0defe-webhook-cert\") pod \"metallb-operator-webhook-server-7df7c8d444-xs67n\" (UID: \"ffe179b8-a1c8-430b-94f5-920aacf0defe\") " pod="metallb-system/metallb-operator-webhook-server-7df7c8d444-xs67n" Jan 29 16:22:44 crc kubenswrapper[4714]: I0129 16:22:44.032897 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q99dn\" (UniqueName: \"kubernetes.io/projected/ffe179b8-a1c8-430b-94f5-920aacf0defe-kube-api-access-q99dn\") pod \"metallb-operator-webhook-server-7df7c8d444-xs67n\" (UID: \"ffe179b8-a1c8-430b-94f5-920aacf0defe\") " pod="metallb-system/metallb-operator-webhook-server-7df7c8d444-xs67n" Jan 29 16:22:44 crc kubenswrapper[4714]: I0129 16:22:44.037703 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ffe179b8-a1c8-430b-94f5-920aacf0defe-webhook-cert\") pod \"metallb-operator-webhook-server-7df7c8d444-xs67n\" (UID: \"ffe179b8-a1c8-430b-94f5-920aacf0defe\") " pod="metallb-system/metallb-operator-webhook-server-7df7c8d444-xs67n" Jan 29 16:22:44 crc kubenswrapper[4714]: I0129 16:22:44.038486 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ffe179b8-a1c8-430b-94f5-920aacf0defe-apiservice-cert\") pod \"metallb-operator-webhook-server-7df7c8d444-xs67n\" (UID: \"ffe179b8-a1c8-430b-94f5-920aacf0defe\") " pod="metallb-system/metallb-operator-webhook-server-7df7c8d444-xs67n" Jan 29 16:22:44 crc kubenswrapper[4714]: I0129 16:22:44.051981 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q99dn\" (UniqueName: \"kubernetes.io/projected/ffe179b8-a1c8-430b-94f5-920aacf0defe-kube-api-access-q99dn\") pod \"metallb-operator-webhook-server-7df7c8d444-xs67n\" (UID: \"ffe179b8-a1c8-430b-94f5-920aacf0defe\") " pod="metallb-system/metallb-operator-webhook-server-7df7c8d444-xs67n" Jan 29 16:22:44 crc kubenswrapper[4714]: I0129 16:22:44.092379 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7df7c8d444-xs67n" Jan 29 16:22:44 crc kubenswrapper[4714]: I0129 16:22:44.116854 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-586b87b897-zpr4q"] Jan 29 16:22:44 crc kubenswrapper[4714]: W0129 16:22:44.136360 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod432a4f98_877c_4f7a_b2b0_ce273a77450a.slice/crio-c4c9bacbb5eb32277a55ef382bf3de55198d38b22f7af753bd79ad92079bf8d7 WatchSource:0}: Error finding container c4c9bacbb5eb32277a55ef382bf3de55198d38b22f7af753bd79ad92079bf8d7: Status 404 returned error can't find the container with id c4c9bacbb5eb32277a55ef382bf3de55198d38b22f7af753bd79ad92079bf8d7 Jan 29 16:22:44 crc kubenswrapper[4714]: I0129 16:22:44.146310 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9ttgj" Jan 29 16:22:44 crc kubenswrapper[4714]: I0129 16:22:44.146399 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9ttgj" Jan 29 16:22:44 crc kubenswrapper[4714]: I0129 16:22:44.330147 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-586b87b897-zpr4q" event={"ID":"432a4f98-877c-4f7a-b2b0-ce273a77450a","Type":"ContainerStarted","Data":"c4c9bacbb5eb32277a55ef382bf3de55198d38b22f7af753bd79ad92079bf8d7"} Jan 29 16:22:44 crc kubenswrapper[4714]: I0129 16:22:44.565944 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7df7c8d444-xs67n"] Jan 29 16:22:44 crc kubenswrapper[4714]: W0129 16:22:44.574966 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffe179b8_a1c8_430b_94f5_920aacf0defe.slice/crio-7d83d11710d4bdc281f2aed567ad9c874b0f2e6cbaee91d3ee9f2f812fe1cd00 WatchSource:0}: Error finding container 7d83d11710d4bdc281f2aed567ad9c874b0f2e6cbaee91d3ee9f2f812fe1cd00: Status 404 returned error can't find the container with id 7d83d11710d4bdc281f2aed567ad9c874b0f2e6cbaee91d3ee9f2f812fe1cd00 Jan 29 16:22:45 crc kubenswrapper[4714]: I0129 16:22:45.196910 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9ttgj" podUID="3b93b136-c182-4b20-89a4-1d61e1d2d03c" containerName="registry-server" probeResult="failure" output=< Jan 29 16:22:45 crc kubenswrapper[4714]: timeout: failed to connect service ":50051" within 1s Jan 29 16:22:45 crc kubenswrapper[4714]: > Jan 29 16:22:45 crc kubenswrapper[4714]: I0129 16:22:45.333905 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7df7c8d444-xs67n" event={"ID":"ffe179b8-a1c8-430b-94f5-920aacf0defe","Type":"ContainerStarted","Data":"7d83d11710d4bdc281f2aed567ad9c874b0f2e6cbaee91d3ee9f2f812fe1cd00"} Jan 29 16:22:54 crc kubenswrapper[4714]: I0129 16:22:54.201037 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9ttgj" Jan 29 16:22:54 crc kubenswrapper[4714]: I0129 16:22:54.256761 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9ttgj" Jan 29 16:22:54 crc kubenswrapper[4714]: I0129 16:22:54.458233 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9ttgj"] Jan 29 16:22:55 crc kubenswrapper[4714]: I0129 16:22:55.409048 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9ttgj" podUID="3b93b136-c182-4b20-89a4-1d61e1d2d03c" containerName="registry-server" containerID="cri-o://71891b59c66ba86ccba3be13ccefa3eee4b9a927ebc16ecd2fd3f3077bf4cab2" gracePeriod=2 Jan 29 16:22:56 crc kubenswrapper[4714]: I0129 16:22:56.422826 4714 generic.go:334] "Generic (PLEG): container finished" podID="3b93b136-c182-4b20-89a4-1d61e1d2d03c" containerID="71891b59c66ba86ccba3be13ccefa3eee4b9a927ebc16ecd2fd3f3077bf4cab2" exitCode=0 Jan 29 16:22:56 crc kubenswrapper[4714]: I0129 16:22:56.422878 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ttgj" event={"ID":"3b93b136-c182-4b20-89a4-1d61e1d2d03c","Type":"ContainerDied","Data":"71891b59c66ba86ccba3be13ccefa3eee4b9a927ebc16ecd2fd3f3077bf4cab2"} Jan 29 16:22:56 crc kubenswrapper[4714]: I0129 16:22:56.758421 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9ttgj" Jan 29 16:22:56 crc kubenswrapper[4714]: I0129 16:22:56.915560 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b93b136-c182-4b20-89a4-1d61e1d2d03c-utilities\") pod \"3b93b136-c182-4b20-89a4-1d61e1d2d03c\" (UID: \"3b93b136-c182-4b20-89a4-1d61e1d2d03c\") " Jan 29 16:22:56 crc kubenswrapper[4714]: I0129 16:22:56.915674 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkw5r\" (UniqueName: \"kubernetes.io/projected/3b93b136-c182-4b20-89a4-1d61e1d2d03c-kube-api-access-xkw5r\") pod \"3b93b136-c182-4b20-89a4-1d61e1d2d03c\" (UID: \"3b93b136-c182-4b20-89a4-1d61e1d2d03c\") " Jan 29 16:22:56 crc kubenswrapper[4714]: I0129 16:22:56.915738 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b93b136-c182-4b20-89a4-1d61e1d2d03c-catalog-content\") pod \"3b93b136-c182-4b20-89a4-1d61e1d2d03c\" (UID: \"3b93b136-c182-4b20-89a4-1d61e1d2d03c\") " Jan 29 16:22:56 crc kubenswrapper[4714]: I0129 16:22:56.918969 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b93b136-c182-4b20-89a4-1d61e1d2d03c-utilities" (OuterVolumeSpecName: "utilities") pod "3b93b136-c182-4b20-89a4-1d61e1d2d03c" (UID: "3b93b136-c182-4b20-89a4-1d61e1d2d03c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:22:56 crc kubenswrapper[4714]: I0129 16:22:56.921216 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b93b136-c182-4b20-89a4-1d61e1d2d03c-kube-api-access-xkw5r" (OuterVolumeSpecName: "kube-api-access-xkw5r") pod "3b93b136-c182-4b20-89a4-1d61e1d2d03c" (UID: "3b93b136-c182-4b20-89a4-1d61e1d2d03c"). InnerVolumeSpecName "kube-api-access-xkw5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:22:57 crc kubenswrapper[4714]: I0129 16:22:57.017876 4714 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b93b136-c182-4b20-89a4-1d61e1d2d03c-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:22:57 crc kubenswrapper[4714]: I0129 16:22:57.017914 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkw5r\" (UniqueName: \"kubernetes.io/projected/3b93b136-c182-4b20-89a4-1d61e1d2d03c-kube-api-access-xkw5r\") on node \"crc\" DevicePath \"\"" Jan 29 16:22:57 crc kubenswrapper[4714]: I0129 16:22:57.044234 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b93b136-c182-4b20-89a4-1d61e1d2d03c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b93b136-c182-4b20-89a4-1d61e1d2d03c" (UID: "3b93b136-c182-4b20-89a4-1d61e1d2d03c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:22:57 crc kubenswrapper[4714]: I0129 16:22:57.119305 4714 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b93b136-c182-4b20-89a4-1d61e1d2d03c-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:22:57 crc kubenswrapper[4714]: I0129 16:22:57.431967 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7df7c8d444-xs67n" event={"ID":"ffe179b8-a1c8-430b-94f5-920aacf0defe","Type":"ContainerStarted","Data":"ebb19f1188d4740707487d9496402b9a5006a1e2b090cbe22f57054f22b1098a"} Jan 29 16:22:57 crc kubenswrapper[4714]: I0129 16:22:57.434928 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ttgj" event={"ID":"3b93b136-c182-4b20-89a4-1d61e1d2d03c","Type":"ContainerDied","Data":"4716263f07ca6d3ad4b5be1c8fb4857b1c649402f03ef208b20007f3c151a261"} Jan 29 16:22:57 crc kubenswrapper[4714]: I0129 16:22:57.435078 4714 scope.go:117] "RemoveContainer" containerID="71891b59c66ba86ccba3be13ccefa3eee4b9a927ebc16ecd2fd3f3077bf4cab2" Jan 29 16:22:57 crc kubenswrapper[4714]: I0129 16:22:57.435276 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9ttgj" Jan 29 16:22:57 crc kubenswrapper[4714]: I0129 16:22:57.445225 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-586b87b897-zpr4q" event={"ID":"432a4f98-877c-4f7a-b2b0-ce273a77450a","Type":"ContainerStarted","Data":"8f2324e58296910aa2839a32886e6fbb898951f4aeaaf7bad1f63e19dd181e23"} Jan 29 16:22:57 crc kubenswrapper[4714]: I0129 16:22:57.458500 4714 scope.go:117] "RemoveContainer" containerID="6632448df0e09a8f0e5148074bcd76d6b08015e4e356eeae9a581a934dfb3cac" Jan 29 16:22:57 crc kubenswrapper[4714]: I0129 16:22:57.495673 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9ttgj"] Jan 29 16:22:57 crc kubenswrapper[4714]: I0129 16:22:57.501051 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9ttgj"] Jan 29 16:22:57 crc kubenswrapper[4714]: I0129 16:22:57.519172 4714 scope.go:117] "RemoveContainer" containerID="dc5891a8c29afb2826ce230ab3fd28a92cd17c413c6d5d58f4c0d944c000e9ef" Jan 29 16:22:57 crc kubenswrapper[4714]: I0129 16:22:57.844404 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:22:57 crc kubenswrapper[4714]: I0129 16:22:57.844472 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:22:58 crc kubenswrapper[4714]: I0129 16:22:58.194836 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b93b136-c182-4b20-89a4-1d61e1d2d03c" path="/var/lib/kubelet/pods/3b93b136-c182-4b20-89a4-1d61e1d2d03c/volumes" Jan 29 16:22:58 crc kubenswrapper[4714]: I0129 16:22:58.455033 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-586b87b897-zpr4q" Jan 29 16:22:58 crc kubenswrapper[4714]: I0129 16:22:58.455113 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7df7c8d444-xs67n" Jan 29 16:22:58 crc kubenswrapper[4714]: I0129 16:22:58.631463 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7df7c8d444-xs67n" podStartSLOduration=3.298912123 podStartE2EDuration="15.6314476s" podCreationTimestamp="2026-01-29 16:22:43 +0000 UTC" firstStartedPulling="2026-01-29 16:22:44.578312013 +0000 UTC m=+771.098813133" lastFinishedPulling="2026-01-29 16:22:56.91084749 +0000 UTC m=+783.431348610" observedRunningTime="2026-01-29 16:22:58.627840338 +0000 UTC m=+785.148341478" watchObservedRunningTime="2026-01-29 16:22:58.6314476 +0000 UTC m=+785.151948720" Jan 29 16:22:58 crc kubenswrapper[4714]: I0129 16:22:58.660797 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-586b87b897-zpr4q" podStartSLOduration=2.916059058 podStartE2EDuration="15.660776292s" podCreationTimestamp="2026-01-29 16:22:43 +0000 UTC" firstStartedPulling="2026-01-29 16:22:44.143196953 +0000 UTC m=+770.663698063" lastFinishedPulling="2026-01-29 16:22:56.887914167 +0000 UTC m=+783.408415297" observedRunningTime="2026-01-29 16:22:58.656557851 +0000 UTC m=+785.177058991" watchObservedRunningTime="2026-01-29 16:22:58.660776292 +0000 UTC m=+785.181277422" Jan 29 16:23:14 crc kubenswrapper[4714]: I0129 16:23:14.096866 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7df7c8d444-xs67n" Jan 29 16:23:27 crc kubenswrapper[4714]: I0129 16:23:27.844532 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:23:27 crc kubenswrapper[4714]: I0129 16:23:27.845073 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:23:33 crc kubenswrapper[4714]: I0129 16:23:33.863483 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-586b87b897-zpr4q" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.643197 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-kk79r"] Jan 29 16:23:34 crc kubenswrapper[4714]: E0129 16:23:34.643672 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b93b136-c182-4b20-89a4-1d61e1d2d03c" containerName="extract-content" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.643689 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b93b136-c182-4b20-89a4-1d61e1d2d03c" containerName="extract-content" Jan 29 16:23:34 crc kubenswrapper[4714]: E0129 16:23:34.643706 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b93b136-c182-4b20-89a4-1d61e1d2d03c" containerName="extract-utilities" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.643713 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b93b136-c182-4b20-89a4-1d61e1d2d03c" containerName="extract-utilities" Jan 29 16:23:34 crc kubenswrapper[4714]: E0129 16:23:34.643722 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b93b136-c182-4b20-89a4-1d61e1d2d03c" containerName="registry-server" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.643728 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b93b136-c182-4b20-89a4-1d61e1d2d03c" containerName="registry-server" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.643820 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b93b136-c182-4b20-89a4-1d61e1d2d03c" containerName="registry-server" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.644176 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-kk79r" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.650305 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-59pmz"] Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.652224 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.661358 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-kk79r"] Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.661980 4714 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-kpj2m" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.662382 4714 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.662671 4714 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.662978 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.737192 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-7mmsh"] Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.738018 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-7mmsh" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.741944 4714 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.741964 4714 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.742152 4714 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-dh498" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.742156 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.749338 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9bbfcf92-8a27-4ba0-9017-7c36906791c8-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-kk79r\" (UID: \"9bbfcf92-8a27-4ba0-9017-7c36906791c8\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-kk79r" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.749378 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a97dd473-5873-4aa1-9166-f7a0c6581be1-metrics-certs\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.749424 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a97dd473-5873-4aa1-9166-f7a0c6581be1-frr-conf\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.749463 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgr67\" (UniqueName: \"kubernetes.io/projected/a97dd473-5873-4aa1-9166-f7a0c6581be1-kube-api-access-bgr67\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.749511 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvvtv\" (UniqueName: \"kubernetes.io/projected/9bbfcf92-8a27-4ba0-9017-7c36906791c8-kube-api-access-cvvtv\") pod \"frr-k8s-webhook-server-7df86c4f6c-kk79r\" (UID: \"9bbfcf92-8a27-4ba0-9017-7c36906791c8\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-kk79r" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.749532 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a97dd473-5873-4aa1-9166-f7a0c6581be1-reloader\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.749617 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a97dd473-5873-4aa1-9166-f7a0c6581be1-metrics\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.749717 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a97dd473-5873-4aa1-9166-f7a0c6581be1-frr-sockets\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.749821 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a97dd473-5873-4aa1-9166-f7a0c6581be1-frr-startup\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.755509 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6968d8fdc4-m26zh"] Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.756277 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-m26zh" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.757958 4714 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.768854 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-m26zh"] Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.851310 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a97dd473-5873-4aa1-9166-f7a0c6581be1-reloader\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.851561 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a97dd473-5873-4aa1-9166-f7a0c6581be1-metrics\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.851646 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a97dd473-5873-4aa1-9166-f7a0c6581be1-frr-sockets\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.851719 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/78b34628-144f-416a-b493-15ba445caa48-cert\") pod \"controller-6968d8fdc4-m26zh\" (UID: \"78b34628-144f-416a-b493-15ba445caa48\") " pod="metallb-system/controller-6968d8fdc4-m26zh" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.851793 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/78b34628-144f-416a-b493-15ba445caa48-metrics-certs\") pod \"controller-6968d8fdc4-m26zh\" (UID: \"78b34628-144f-416a-b493-15ba445caa48\") " pod="metallb-system/controller-6968d8fdc4-m26zh" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.851870 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k9wr\" (UniqueName: \"kubernetes.io/projected/813f735d-8336-49e9-b018-e6dbf74ddc99-kube-api-access-2k9wr\") pod \"speaker-7mmsh\" (UID: \"813f735d-8336-49e9-b018-e6dbf74ddc99\") " pod="metallb-system/speaker-7mmsh" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.851821 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a97dd473-5873-4aa1-9166-f7a0c6581be1-reloader\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.851967 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a97dd473-5873-4aa1-9166-f7a0c6581be1-frr-sockets\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.852041 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/813f735d-8336-49e9-b018-e6dbf74ddc99-memberlist\") pod \"speaker-7mmsh\" (UID: \"813f735d-8336-49e9-b018-e6dbf74ddc99\") " pod="metallb-system/speaker-7mmsh" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.852115 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kjbj\" (UniqueName: \"kubernetes.io/projected/78b34628-144f-416a-b493-15ba445caa48-kube-api-access-5kjbj\") pod \"controller-6968d8fdc4-m26zh\" (UID: \"78b34628-144f-416a-b493-15ba445caa48\") " pod="metallb-system/controller-6968d8fdc4-m26zh" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.852185 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a97dd473-5873-4aa1-9166-f7a0c6581be1-frr-startup\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.852273 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9bbfcf92-8a27-4ba0-9017-7c36906791c8-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-kk79r\" (UID: \"9bbfcf92-8a27-4ba0-9017-7c36906791c8\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-kk79r" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.852338 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a97dd473-5873-4aa1-9166-f7a0c6581be1-metrics-certs\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.852226 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a97dd473-5873-4aa1-9166-f7a0c6581be1-metrics\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.852460 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/813f735d-8336-49e9-b018-e6dbf74ddc99-metrics-certs\") pod \"speaker-7mmsh\" (UID: \"813f735d-8336-49e9-b018-e6dbf74ddc99\") " pod="metallb-system/speaker-7mmsh" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.852544 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a97dd473-5873-4aa1-9166-f7a0c6581be1-frr-conf\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.852624 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgr67\" (UniqueName: \"kubernetes.io/projected/a97dd473-5873-4aa1-9166-f7a0c6581be1-kube-api-access-bgr67\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.852700 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/813f735d-8336-49e9-b018-e6dbf74ddc99-metallb-excludel2\") pod \"speaker-7mmsh\" (UID: \"813f735d-8336-49e9-b018-e6dbf74ddc99\") " pod="metallb-system/speaker-7mmsh" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.852799 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvvtv\" (UniqueName: \"kubernetes.io/projected/9bbfcf92-8a27-4ba0-9017-7c36906791c8-kube-api-access-cvvtv\") pod \"frr-k8s-webhook-server-7df86c4f6c-kk79r\" (UID: \"9bbfcf92-8a27-4ba0-9017-7c36906791c8\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-kk79r" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.852802 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a97dd473-5873-4aa1-9166-f7a0c6581be1-frr-conf\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.853502 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a97dd473-5873-4aa1-9166-f7a0c6581be1-frr-startup\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.861595 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a97dd473-5873-4aa1-9166-f7a0c6581be1-metrics-certs\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.862332 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9bbfcf92-8a27-4ba0-9017-7c36906791c8-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-kk79r\" (UID: \"9bbfcf92-8a27-4ba0-9017-7c36906791c8\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-kk79r" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.867801 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvvtv\" (UniqueName: \"kubernetes.io/projected/9bbfcf92-8a27-4ba0-9017-7c36906791c8-kube-api-access-cvvtv\") pod \"frr-k8s-webhook-server-7df86c4f6c-kk79r\" (UID: \"9bbfcf92-8a27-4ba0-9017-7c36906791c8\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-kk79r" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.868595 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgr67\" (UniqueName: \"kubernetes.io/projected/a97dd473-5873-4aa1-9166-f7a0c6581be1-kube-api-access-bgr67\") pod \"frr-k8s-59pmz\" (UID: \"a97dd473-5873-4aa1-9166-f7a0c6581be1\") " pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.954440 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/813f735d-8336-49e9-b018-e6dbf74ddc99-metrics-certs\") pod \"speaker-7mmsh\" (UID: \"813f735d-8336-49e9-b018-e6dbf74ddc99\") " pod="metallb-system/speaker-7mmsh" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.954503 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/813f735d-8336-49e9-b018-e6dbf74ddc99-metallb-excludel2\") pod \"speaker-7mmsh\" (UID: \"813f735d-8336-49e9-b018-e6dbf74ddc99\") " pod="metallb-system/speaker-7mmsh" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.954541 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/78b34628-144f-416a-b493-15ba445caa48-cert\") pod \"controller-6968d8fdc4-m26zh\" (UID: \"78b34628-144f-416a-b493-15ba445caa48\") " pod="metallb-system/controller-6968d8fdc4-m26zh" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.954564 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/78b34628-144f-416a-b493-15ba445caa48-metrics-certs\") pod \"controller-6968d8fdc4-m26zh\" (UID: \"78b34628-144f-416a-b493-15ba445caa48\") " pod="metallb-system/controller-6968d8fdc4-m26zh" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.954586 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k9wr\" (UniqueName: \"kubernetes.io/projected/813f735d-8336-49e9-b018-e6dbf74ddc99-kube-api-access-2k9wr\") pod \"speaker-7mmsh\" (UID: \"813f735d-8336-49e9-b018-e6dbf74ddc99\") " pod="metallb-system/speaker-7mmsh" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.954615 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/813f735d-8336-49e9-b018-e6dbf74ddc99-memberlist\") pod \"speaker-7mmsh\" (UID: \"813f735d-8336-49e9-b018-e6dbf74ddc99\") " pod="metallb-system/speaker-7mmsh" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.954636 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kjbj\" (UniqueName: \"kubernetes.io/projected/78b34628-144f-416a-b493-15ba445caa48-kube-api-access-5kjbj\") pod \"controller-6968d8fdc4-m26zh\" (UID: \"78b34628-144f-416a-b493-15ba445caa48\") " pod="metallb-system/controller-6968d8fdc4-m26zh" Jan 29 16:23:34 crc kubenswrapper[4714]: E0129 16:23:34.955047 4714 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 29 16:23:34 crc kubenswrapper[4714]: E0129 16:23:34.955098 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/813f735d-8336-49e9-b018-e6dbf74ddc99-memberlist podName:813f735d-8336-49e9-b018-e6dbf74ddc99 nodeName:}" failed. No retries permitted until 2026-01-29 16:23:35.455081671 +0000 UTC m=+821.975582791 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/813f735d-8336-49e9-b018-e6dbf74ddc99-memberlist") pod "speaker-7mmsh" (UID: "813f735d-8336-49e9-b018-e6dbf74ddc99") : secret "metallb-memberlist" not found Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.955816 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/813f735d-8336-49e9-b018-e6dbf74ddc99-metallb-excludel2\") pod \"speaker-7mmsh\" (UID: \"813f735d-8336-49e9-b018-e6dbf74ddc99\") " pod="metallb-system/speaker-7mmsh" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.957212 4714 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.959526 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/813f735d-8336-49e9-b018-e6dbf74ddc99-metrics-certs\") pod \"speaker-7mmsh\" (UID: \"813f735d-8336-49e9-b018-e6dbf74ddc99\") " pod="metallb-system/speaker-7mmsh" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.960010 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/78b34628-144f-416a-b493-15ba445caa48-metrics-certs\") pod \"controller-6968d8fdc4-m26zh\" (UID: \"78b34628-144f-416a-b493-15ba445caa48\") " pod="metallb-system/controller-6968d8fdc4-m26zh" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.967043 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-kk79r" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.973871 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k9wr\" (UniqueName: \"kubernetes.io/projected/813f735d-8336-49e9-b018-e6dbf74ddc99-kube-api-access-2k9wr\") pod \"speaker-7mmsh\" (UID: \"813f735d-8336-49e9-b018-e6dbf74ddc99\") " pod="metallb-system/speaker-7mmsh" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.975199 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kjbj\" (UniqueName: \"kubernetes.io/projected/78b34628-144f-416a-b493-15ba445caa48-kube-api-access-5kjbj\") pod \"controller-6968d8fdc4-m26zh\" (UID: \"78b34628-144f-416a-b493-15ba445caa48\") " pod="metallb-system/controller-6968d8fdc4-m26zh" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.975608 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:34 crc kubenswrapper[4714]: I0129 16:23:34.976289 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/78b34628-144f-416a-b493-15ba445caa48-cert\") pod \"controller-6968d8fdc4-m26zh\" (UID: \"78b34628-144f-416a-b493-15ba445caa48\") " pod="metallb-system/controller-6968d8fdc4-m26zh" Jan 29 16:23:35 crc kubenswrapper[4714]: I0129 16:23:35.072161 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-m26zh" Jan 29 16:23:35 crc kubenswrapper[4714]: I0129 16:23:35.328255 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-m26zh"] Jan 29 16:23:35 crc kubenswrapper[4714]: I0129 16:23:35.462318 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/813f735d-8336-49e9-b018-e6dbf74ddc99-memberlist\") pod \"speaker-7mmsh\" (UID: \"813f735d-8336-49e9-b018-e6dbf74ddc99\") " pod="metallb-system/speaker-7mmsh" Jan 29 16:23:35 crc kubenswrapper[4714]: E0129 16:23:35.462490 4714 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 29 16:23:35 crc kubenswrapper[4714]: E0129 16:23:35.462564 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/813f735d-8336-49e9-b018-e6dbf74ddc99-memberlist podName:813f735d-8336-49e9-b018-e6dbf74ddc99 nodeName:}" failed. No retries permitted until 2026-01-29 16:23:36.462549627 +0000 UTC m=+822.983050747 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/813f735d-8336-49e9-b018-e6dbf74ddc99-memberlist") pod "speaker-7mmsh" (UID: "813f735d-8336-49e9-b018-e6dbf74ddc99") : secret "metallb-memberlist" not found Jan 29 16:23:35 crc kubenswrapper[4714]: I0129 16:23:35.469384 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-kk79r"] Jan 29 16:23:35 crc kubenswrapper[4714]: W0129 16:23:35.473345 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9bbfcf92_8a27_4ba0_9017_7c36906791c8.slice/crio-eb293dcb532fa6a0e844c468662976cc97b3ffe233e3d39cbd3ad550635645a0 WatchSource:0}: Error finding container eb293dcb532fa6a0e844c468662976cc97b3ffe233e3d39cbd3ad550635645a0: Status 404 returned error can't find the container with id eb293dcb532fa6a0e844c468662976cc97b3ffe233e3d39cbd3ad550635645a0 Jan 29 16:23:35 crc kubenswrapper[4714]: I0129 16:23:35.682082 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-kk79r" event={"ID":"9bbfcf92-8a27-4ba0-9017-7c36906791c8","Type":"ContainerStarted","Data":"eb293dcb532fa6a0e844c468662976cc97b3ffe233e3d39cbd3ad550635645a0"} Jan 29 16:23:35 crc kubenswrapper[4714]: I0129 16:23:35.683846 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-m26zh" event={"ID":"78b34628-144f-416a-b493-15ba445caa48","Type":"ContainerStarted","Data":"42678d585a93ee4e807989c839b6083070ee10ae0eaf4c398d2cbf846d9e2bf9"} Jan 29 16:23:35 crc kubenswrapper[4714]: I0129 16:23:35.683880 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-m26zh" event={"ID":"78b34628-144f-416a-b493-15ba445caa48","Type":"ContainerStarted","Data":"d65afc9c64b6337e908d8e47d9121e80cfa91d9056cb7eb9f807063d779583ac"} Jan 29 16:23:35 crc kubenswrapper[4714]: I0129 16:23:35.684796 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-59pmz" event={"ID":"a97dd473-5873-4aa1-9166-f7a0c6581be1","Type":"ContainerStarted","Data":"54e1ecd7827f449b3ac55f819a0ff9fe56bcf3d7e18090da24e66a24fe177b4c"} Jan 29 16:23:36 crc kubenswrapper[4714]: I0129 16:23:36.477022 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/813f735d-8336-49e9-b018-e6dbf74ddc99-memberlist\") pod \"speaker-7mmsh\" (UID: \"813f735d-8336-49e9-b018-e6dbf74ddc99\") " pod="metallb-system/speaker-7mmsh" Jan 29 16:23:36 crc kubenswrapper[4714]: I0129 16:23:36.482717 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/813f735d-8336-49e9-b018-e6dbf74ddc99-memberlist\") pod \"speaker-7mmsh\" (UID: \"813f735d-8336-49e9-b018-e6dbf74ddc99\") " pod="metallb-system/speaker-7mmsh" Jan 29 16:23:36 crc kubenswrapper[4714]: I0129 16:23:36.551175 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-7mmsh" Jan 29 16:23:36 crc kubenswrapper[4714]: I0129 16:23:36.692543 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7mmsh" event={"ID":"813f735d-8336-49e9-b018-e6dbf74ddc99","Type":"ContainerStarted","Data":"1d545643ff622961dc3f0863b6328e6e18150121c23013cbe42045322f9e36f3"} Jan 29 16:23:37 crc kubenswrapper[4714]: I0129 16:23:37.705787 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7mmsh" event={"ID":"813f735d-8336-49e9-b018-e6dbf74ddc99","Type":"ContainerStarted","Data":"aaa70c96c3fba2bd7a53fe7ad0d780d5d36803534760758c3d9bfc9dd4187178"} Jan 29 16:23:39 crc kubenswrapper[4714]: I0129 16:23:39.717278 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-m26zh" event={"ID":"78b34628-144f-416a-b493-15ba445caa48","Type":"ContainerStarted","Data":"c3d0e7f3954182f3b505afdfd5742def976e5a1df65e4d6a4dd928cfc467edf4"} Jan 29 16:23:39 crc kubenswrapper[4714]: I0129 16:23:39.717633 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6968d8fdc4-m26zh" Jan 29 16:23:39 crc kubenswrapper[4714]: I0129 16:23:39.719781 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7mmsh" event={"ID":"813f735d-8336-49e9-b018-e6dbf74ddc99","Type":"ContainerStarted","Data":"6aba26b9b5b0b50b75e9b9e6ceda106719062ca4e6121e67474c51352a428a90"} Jan 29 16:23:39 crc kubenswrapper[4714]: I0129 16:23:39.719958 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-7mmsh" Jan 29 16:23:39 crc kubenswrapper[4714]: I0129 16:23:39.736998 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6968d8fdc4-m26zh" podStartSLOduration=2.136871459 podStartE2EDuration="5.73697936s" podCreationTimestamp="2026-01-29 16:23:34 +0000 UTC" firstStartedPulling="2026-01-29 16:23:35.429802619 +0000 UTC m=+821.950303739" lastFinishedPulling="2026-01-29 16:23:39.02991052 +0000 UTC m=+825.550411640" observedRunningTime="2026-01-29 16:23:39.733837733 +0000 UTC m=+826.254338853" watchObservedRunningTime="2026-01-29 16:23:39.73697936 +0000 UTC m=+826.257480480" Jan 29 16:23:39 crc kubenswrapper[4714]: I0129 16:23:39.758580 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-7mmsh" podStartSLOduration=3.646873782 podStartE2EDuration="5.758566501s" podCreationTimestamp="2026-01-29 16:23:34 +0000 UTC" firstStartedPulling="2026-01-29 16:23:36.93685673 +0000 UTC m=+823.457357850" lastFinishedPulling="2026-01-29 16:23:39.048549449 +0000 UTC m=+825.569050569" observedRunningTime="2026-01-29 16:23:39.758429757 +0000 UTC m=+826.278930887" watchObservedRunningTime="2026-01-29 16:23:39.758566501 +0000 UTC m=+826.279067621" Jan 29 16:23:42 crc kubenswrapper[4714]: I0129 16:23:42.736977 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-kk79r" event={"ID":"9bbfcf92-8a27-4ba0-9017-7c36906791c8","Type":"ContainerStarted","Data":"df1c444d10af60503ec695ecd526e96e371ce9d52cac5487bf24f6a493a8956d"} Jan 29 16:23:42 crc kubenswrapper[4714]: I0129 16:23:42.737348 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-kk79r" Jan 29 16:23:42 crc kubenswrapper[4714]: I0129 16:23:42.739539 4714 generic.go:334] "Generic (PLEG): container finished" podID="a97dd473-5873-4aa1-9166-f7a0c6581be1" containerID="e8eb896c49334701cb329d19d54e164c9ae10325bd6a724e508b54dbaad17ad0" exitCode=0 Jan 29 16:23:42 crc kubenswrapper[4714]: I0129 16:23:42.739593 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-59pmz" event={"ID":"a97dd473-5873-4aa1-9166-f7a0c6581be1","Type":"ContainerDied","Data":"e8eb896c49334701cb329d19d54e164c9ae10325bd6a724e508b54dbaad17ad0"} Jan 29 16:23:42 crc kubenswrapper[4714]: I0129 16:23:42.760311 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-kk79r" podStartSLOduration=1.71547556 podStartE2EDuration="8.760279489s" podCreationTimestamp="2026-01-29 16:23:34 +0000 UTC" firstStartedPulling="2026-01-29 16:23:35.476622395 +0000 UTC m=+821.997123515" lastFinishedPulling="2026-01-29 16:23:42.521426314 +0000 UTC m=+829.041927444" observedRunningTime="2026-01-29 16:23:42.756689967 +0000 UTC m=+829.277191117" watchObservedRunningTime="2026-01-29 16:23:42.760279489 +0000 UTC m=+829.280780669" Jan 29 16:23:43 crc kubenswrapper[4714]: I0129 16:23:43.750057 4714 generic.go:334] "Generic (PLEG): container finished" podID="a97dd473-5873-4aa1-9166-f7a0c6581be1" containerID="bb039282043c0e2e27e8fbb3aaa72038d17c43384473549fe668defe9937c3bd" exitCode=0 Jan 29 16:23:43 crc kubenswrapper[4714]: I0129 16:23:43.750204 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-59pmz" event={"ID":"a97dd473-5873-4aa1-9166-f7a0c6581be1","Type":"ContainerDied","Data":"bb039282043c0e2e27e8fbb3aaa72038d17c43384473549fe668defe9937c3bd"} Jan 29 16:23:44 crc kubenswrapper[4714]: I0129 16:23:44.759145 4714 generic.go:334] "Generic (PLEG): container finished" podID="a97dd473-5873-4aa1-9166-f7a0c6581be1" containerID="0d159d7890f1bc21ee4bf02da74cfc8ebcc219bc17c61ccca4fb06b34bd89880" exitCode=0 Jan 29 16:23:44 crc kubenswrapper[4714]: I0129 16:23:44.759303 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-59pmz" event={"ID":"a97dd473-5873-4aa1-9166-f7a0c6581be1","Type":"ContainerDied","Data":"0d159d7890f1bc21ee4bf02da74cfc8ebcc219bc17c61ccca4fb06b34bd89880"} Jan 29 16:23:45 crc kubenswrapper[4714]: I0129 16:23:45.090143 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6968d8fdc4-m26zh" Jan 29 16:23:45 crc kubenswrapper[4714]: I0129 16:23:45.774997 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-59pmz" event={"ID":"a97dd473-5873-4aa1-9166-f7a0c6581be1","Type":"ContainerStarted","Data":"e6b47835a151d644b4b184747c5be8a671e14885f81ea0d64c4a2f8066c1c93d"} Jan 29 16:23:45 crc kubenswrapper[4714]: I0129 16:23:45.775063 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-59pmz" event={"ID":"a97dd473-5873-4aa1-9166-f7a0c6581be1","Type":"ContainerStarted","Data":"42f637cc9a9cb1a567464df9bf350b9575f31b124eb87d2a9f41faf4265d58ce"} Jan 29 16:23:45 crc kubenswrapper[4714]: I0129 16:23:45.775079 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-59pmz" event={"ID":"a97dd473-5873-4aa1-9166-f7a0c6581be1","Type":"ContainerStarted","Data":"e4eb656ef2f36a80ede4736e536c5a4d89016f3180165779fe81680c7e88dda5"} Jan 29 16:23:45 crc kubenswrapper[4714]: I0129 16:23:45.775092 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-59pmz" event={"ID":"a97dd473-5873-4aa1-9166-f7a0c6581be1","Type":"ContainerStarted","Data":"b86c31da87199869693e3afa55f5db0defdbe590c0ca1e39d3d2937bacc30075"} Jan 29 16:23:45 crc kubenswrapper[4714]: I0129 16:23:45.775104 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-59pmz" event={"ID":"a97dd473-5873-4aa1-9166-f7a0c6581be1","Type":"ContainerStarted","Data":"896a7e5790b8e818708c993aca3b3cacb73f13ca0a6340e7992bf210ae77e6c7"} Jan 29 16:23:46 crc kubenswrapper[4714]: I0129 16:23:46.556904 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-7mmsh" Jan 29 16:23:46 crc kubenswrapper[4714]: I0129 16:23:46.792578 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-59pmz" event={"ID":"a97dd473-5873-4aa1-9166-f7a0c6581be1","Type":"ContainerStarted","Data":"9168595a16c083d33f31bdc99300f584dd37b304de01c3c16e0c277103745a6e"} Jan 29 16:23:46 crc kubenswrapper[4714]: I0129 16:23:46.792923 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:46 crc kubenswrapper[4714]: I0129 16:23:46.836042 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-59pmz" podStartSLOduration=5.49155512 podStartE2EDuration="12.836022306s" podCreationTimestamp="2026-01-29 16:23:34 +0000 UTC" firstStartedPulling="2026-01-29 16:23:35.157647228 +0000 UTC m=+821.678148348" lastFinishedPulling="2026-01-29 16:23:42.502114414 +0000 UTC m=+829.022615534" observedRunningTime="2026-01-29 16:23:46.834169008 +0000 UTC m=+833.354670188" watchObservedRunningTime="2026-01-29 16:23:46.836022306 +0000 UTC m=+833.356523426" Jan 29 16:23:49 crc kubenswrapper[4714]: I0129 16:23:49.976774 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:50 crc kubenswrapper[4714]: I0129 16:23:50.046594 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:52 crc kubenswrapper[4714]: I0129 16:23:52.526144 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-jxfdk"] Jan 29 16:23:52 crc kubenswrapper[4714]: I0129 16:23:52.527197 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-jxfdk" Jan 29 16:23:52 crc kubenswrapper[4714]: I0129 16:23:52.530645 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-zfxjm" Jan 29 16:23:52 crc kubenswrapper[4714]: I0129 16:23:52.530877 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Jan 29 16:23:52 crc kubenswrapper[4714]: I0129 16:23:52.531015 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Jan 29 16:23:52 crc kubenswrapper[4714]: I0129 16:23:52.592671 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-jxfdk"] Jan 29 16:23:52 crc kubenswrapper[4714]: I0129 16:23:52.631035 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtkk7\" (UniqueName: \"kubernetes.io/projected/c73614a5-aed3-4942-9382-2981e22773ec-kube-api-access-wtkk7\") pod \"mariadb-operator-index-jxfdk\" (UID: \"c73614a5-aed3-4942-9382-2981e22773ec\") " pod="openstack-operators/mariadb-operator-index-jxfdk" Jan 29 16:23:52 crc kubenswrapper[4714]: I0129 16:23:52.731855 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtkk7\" (UniqueName: \"kubernetes.io/projected/c73614a5-aed3-4942-9382-2981e22773ec-kube-api-access-wtkk7\") pod \"mariadb-operator-index-jxfdk\" (UID: \"c73614a5-aed3-4942-9382-2981e22773ec\") " pod="openstack-operators/mariadb-operator-index-jxfdk" Jan 29 16:23:52 crc kubenswrapper[4714]: I0129 16:23:52.751858 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtkk7\" (UniqueName: \"kubernetes.io/projected/c73614a5-aed3-4942-9382-2981e22773ec-kube-api-access-wtkk7\") pod \"mariadb-operator-index-jxfdk\" (UID: \"c73614a5-aed3-4942-9382-2981e22773ec\") " pod="openstack-operators/mariadb-operator-index-jxfdk" Jan 29 16:23:52 crc kubenswrapper[4714]: I0129 16:23:52.846678 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-jxfdk" Jan 29 16:23:53 crc kubenswrapper[4714]: I0129 16:23:53.018698 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-jxfdk"] Jan 29 16:23:53 crc kubenswrapper[4714]: W0129 16:23:53.026780 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc73614a5_aed3_4942_9382_2981e22773ec.slice/crio-590a2584ed40d70665967956dedd365e20ebaf918ed6a882e280c2a862bb3df9 WatchSource:0}: Error finding container 590a2584ed40d70665967956dedd365e20ebaf918ed6a882e280c2a862bb3df9: Status 404 returned error can't find the container with id 590a2584ed40d70665967956dedd365e20ebaf918ed6a882e280c2a862bb3df9 Jan 29 16:23:53 crc kubenswrapper[4714]: I0129 16:23:53.837603 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-jxfdk" event={"ID":"c73614a5-aed3-4942-9382-2981e22773ec","Type":"ContainerStarted","Data":"590a2584ed40d70665967956dedd365e20ebaf918ed6a882e280c2a862bb3df9"} Jan 29 16:23:54 crc kubenswrapper[4714]: I0129 16:23:54.846704 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-jxfdk" event={"ID":"c73614a5-aed3-4942-9382-2981e22773ec","Type":"ContainerStarted","Data":"e13c954e2ceec860ed151e1275323a23d1c2600cd51f3926408e9e8970fbca50"} Jan 29 16:23:54 crc kubenswrapper[4714]: I0129 16:23:54.866693 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-jxfdk" podStartSLOduration=1.963355862 podStartE2EDuration="2.866675844s" podCreationTimestamp="2026-01-29 16:23:52 +0000 UTC" firstStartedPulling="2026-01-29 16:23:53.030441269 +0000 UTC m=+839.550942389" lastFinishedPulling="2026-01-29 16:23:53.933761251 +0000 UTC m=+840.454262371" observedRunningTime="2026-01-29 16:23:54.865537328 +0000 UTC m=+841.386038458" watchObservedRunningTime="2026-01-29 16:23:54.866675844 +0000 UTC m=+841.387176954" Jan 29 16:23:54 crc kubenswrapper[4714]: I0129 16:23:54.975576 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-kk79r" Jan 29 16:23:54 crc kubenswrapper[4714]: I0129 16:23:54.978766 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-59pmz" Jan 29 16:23:55 crc kubenswrapper[4714]: I0129 16:23:55.911220 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-jxfdk"] Jan 29 16:23:56 crc kubenswrapper[4714]: I0129 16:23:56.518771 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-l6dkm"] Jan 29 16:23:56 crc kubenswrapper[4714]: I0129 16:23:56.519887 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-l6dkm" Jan 29 16:23:56 crc kubenswrapper[4714]: I0129 16:23:56.532508 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-l6dkm"] Jan 29 16:23:56 crc kubenswrapper[4714]: I0129 16:23:56.688993 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r4sv\" (UniqueName: \"kubernetes.io/projected/d990dfb7-e078-4c7e-8e98-40b10f062a04-kube-api-access-8r4sv\") pod \"mariadb-operator-index-l6dkm\" (UID: \"d990dfb7-e078-4c7e-8e98-40b10f062a04\") " pod="openstack-operators/mariadb-operator-index-l6dkm" Jan 29 16:23:56 crc kubenswrapper[4714]: I0129 16:23:56.790526 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r4sv\" (UniqueName: \"kubernetes.io/projected/d990dfb7-e078-4c7e-8e98-40b10f062a04-kube-api-access-8r4sv\") pod \"mariadb-operator-index-l6dkm\" (UID: \"d990dfb7-e078-4c7e-8e98-40b10f062a04\") " pod="openstack-operators/mariadb-operator-index-l6dkm" Jan 29 16:23:56 crc kubenswrapper[4714]: I0129 16:23:56.816682 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r4sv\" (UniqueName: \"kubernetes.io/projected/d990dfb7-e078-4c7e-8e98-40b10f062a04-kube-api-access-8r4sv\") pod \"mariadb-operator-index-l6dkm\" (UID: \"d990dfb7-e078-4c7e-8e98-40b10f062a04\") " pod="openstack-operators/mariadb-operator-index-l6dkm" Jan 29 16:23:56 crc kubenswrapper[4714]: I0129 16:23:56.856694 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-l6dkm" Jan 29 16:23:56 crc kubenswrapper[4714]: I0129 16:23:56.859437 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-jxfdk" podUID="c73614a5-aed3-4942-9382-2981e22773ec" containerName="registry-server" containerID="cri-o://e13c954e2ceec860ed151e1275323a23d1c2600cd51f3926408e9e8970fbca50" gracePeriod=2 Jan 29 16:23:57 crc kubenswrapper[4714]: I0129 16:23:57.270904 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-l6dkm"] Jan 29 16:23:57 crc kubenswrapper[4714]: W0129 16:23:57.277240 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd990dfb7_e078_4c7e_8e98_40b10f062a04.slice/crio-cd19708e51ad0ae38749b06c635286e93f2554a3fecdeb70a36c4d4f40376c94 WatchSource:0}: Error finding container cd19708e51ad0ae38749b06c635286e93f2554a3fecdeb70a36c4d4f40376c94: Status 404 returned error can't find the container with id cd19708e51ad0ae38749b06c635286e93f2554a3fecdeb70a36c4d4f40376c94 Jan 29 16:23:57 crc kubenswrapper[4714]: I0129 16:23:57.844693 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:23:57 crc kubenswrapper[4714]: I0129 16:23:57.844763 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:23:57 crc kubenswrapper[4714]: I0129 16:23:57.844829 4714 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:23:57 crc kubenswrapper[4714]: I0129 16:23:57.845434 4714 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"434181b332ad91829c9ca3b07c475cac7d3c8b013492e90ce07fd88776d24efa"} pod="openshift-machine-config-operator/machine-config-daemon-ppngk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 16:23:57 crc kubenswrapper[4714]: I0129 16:23:57.845510 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" containerID="cri-o://434181b332ad91829c9ca3b07c475cac7d3c8b013492e90ce07fd88776d24efa" gracePeriod=600 Jan 29 16:23:57 crc kubenswrapper[4714]: I0129 16:23:57.866564 4714 generic.go:334] "Generic (PLEG): container finished" podID="c73614a5-aed3-4942-9382-2981e22773ec" containerID="e13c954e2ceec860ed151e1275323a23d1c2600cd51f3926408e9e8970fbca50" exitCode=0 Jan 29 16:23:57 crc kubenswrapper[4714]: I0129 16:23:57.866656 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-jxfdk" event={"ID":"c73614a5-aed3-4942-9382-2981e22773ec","Type":"ContainerDied","Data":"e13c954e2ceec860ed151e1275323a23d1c2600cd51f3926408e9e8970fbca50"} Jan 29 16:23:57 crc kubenswrapper[4714]: I0129 16:23:57.868457 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-l6dkm" event={"ID":"d990dfb7-e078-4c7e-8e98-40b10f062a04","Type":"ContainerStarted","Data":"cd19708e51ad0ae38749b06c635286e93f2554a3fecdeb70a36c4d4f40376c94"} Jan 29 16:23:58 crc kubenswrapper[4714]: I0129 16:23:58.880120 4714 generic.go:334] "Generic (PLEG): container finished" podID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerID="434181b332ad91829c9ca3b07c475cac7d3c8b013492e90ce07fd88776d24efa" exitCode=0 Jan 29 16:23:58 crc kubenswrapper[4714]: I0129 16:23:58.880173 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" event={"ID":"c8c765f3-89eb-4077-8829-03e86eb0c90c","Type":"ContainerDied","Data":"434181b332ad91829c9ca3b07c475cac7d3c8b013492e90ce07fd88776d24efa"} Jan 29 16:23:58 crc kubenswrapper[4714]: I0129 16:23:58.880212 4714 scope.go:117] "RemoveContainer" containerID="aeda778ca6de188bfb9f09408c5d355e6f8d4366d5f9ebe7bfd9f2e4dea2a0e4" Jan 29 16:23:59 crc kubenswrapper[4714]: I0129 16:23:59.614372 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-jxfdk" Jan 29 16:23:59 crc kubenswrapper[4714]: I0129 16:23:59.728437 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtkk7\" (UniqueName: \"kubernetes.io/projected/c73614a5-aed3-4942-9382-2981e22773ec-kube-api-access-wtkk7\") pod \"c73614a5-aed3-4942-9382-2981e22773ec\" (UID: \"c73614a5-aed3-4942-9382-2981e22773ec\") " Jan 29 16:23:59 crc kubenswrapper[4714]: I0129 16:23:59.733684 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c73614a5-aed3-4942-9382-2981e22773ec-kube-api-access-wtkk7" (OuterVolumeSpecName: "kube-api-access-wtkk7") pod "c73614a5-aed3-4942-9382-2981e22773ec" (UID: "c73614a5-aed3-4942-9382-2981e22773ec"). InnerVolumeSpecName "kube-api-access-wtkk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:23:59 crc kubenswrapper[4714]: I0129 16:23:59.830204 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtkk7\" (UniqueName: \"kubernetes.io/projected/c73614a5-aed3-4942-9382-2981e22773ec-kube-api-access-wtkk7\") on node \"crc\" DevicePath \"\"" Jan 29 16:23:59 crc kubenswrapper[4714]: I0129 16:23:59.888114 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-jxfdk" event={"ID":"c73614a5-aed3-4942-9382-2981e22773ec","Type":"ContainerDied","Data":"590a2584ed40d70665967956dedd365e20ebaf918ed6a882e280c2a862bb3df9"} Jan 29 16:23:59 crc kubenswrapper[4714]: I0129 16:23:59.888162 4714 scope.go:117] "RemoveContainer" containerID="e13c954e2ceec860ed151e1275323a23d1c2600cd51f3926408e9e8970fbca50" Jan 29 16:23:59 crc kubenswrapper[4714]: I0129 16:23:59.888171 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-jxfdk" Jan 29 16:23:59 crc kubenswrapper[4714]: I0129 16:23:59.893294 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" event={"ID":"c8c765f3-89eb-4077-8829-03e86eb0c90c","Type":"ContainerStarted","Data":"77045db0ac9dbee23fe648e58207222e15e50d5178fcc5cc7a606b4bbe2af7ec"} Jan 29 16:23:59 crc kubenswrapper[4714]: I0129 16:23:59.897614 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-l6dkm" event={"ID":"d990dfb7-e078-4c7e-8e98-40b10f062a04","Type":"ContainerStarted","Data":"0020667ef371fcb5a3d00febffc3770f6cb20130544e17735a3ccff225db36b3"} Jan 29 16:23:59 crc kubenswrapper[4714]: I0129 16:23:59.932652 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-l6dkm" podStartSLOduration=1.621026461 podStartE2EDuration="3.932630194s" podCreationTimestamp="2026-01-29 16:23:56 +0000 UTC" firstStartedPulling="2026-01-29 16:23:57.28068193 +0000 UTC m=+843.801183050" lastFinishedPulling="2026-01-29 16:23:59.592285663 +0000 UTC m=+846.112786783" observedRunningTime="2026-01-29 16:23:59.924365537 +0000 UTC m=+846.444866677" watchObservedRunningTime="2026-01-29 16:23:59.932630194 +0000 UTC m=+846.453131314" Jan 29 16:23:59 crc kubenswrapper[4714]: I0129 16:23:59.937960 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-jxfdk"] Jan 29 16:23:59 crc kubenswrapper[4714]: I0129 16:23:59.941080 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-jxfdk"] Jan 29 16:24:00 crc kubenswrapper[4714]: I0129 16:24:00.190952 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c73614a5-aed3-4942-9382-2981e22773ec" path="/var/lib/kubelet/pods/c73614a5-aed3-4942-9382-2981e22773ec/volumes" Jan 29 16:24:06 crc kubenswrapper[4714]: I0129 16:24:06.857311 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-l6dkm" Jan 29 16:24:06 crc kubenswrapper[4714]: I0129 16:24:06.857753 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-l6dkm" Jan 29 16:24:06 crc kubenswrapper[4714]: I0129 16:24:06.886819 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-l6dkm" Jan 29 16:24:06 crc kubenswrapper[4714]: I0129 16:24:06.976573 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-l6dkm" Jan 29 16:24:08 crc kubenswrapper[4714]: I0129 16:24:08.553278 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp"] Jan 29 16:24:08 crc kubenswrapper[4714]: E0129 16:24:08.556615 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c73614a5-aed3-4942-9382-2981e22773ec" containerName="registry-server" Jan 29 16:24:08 crc kubenswrapper[4714]: I0129 16:24:08.556711 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="c73614a5-aed3-4942-9382-2981e22773ec" containerName="registry-server" Jan 29 16:24:08 crc kubenswrapper[4714]: I0129 16:24:08.557141 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="c73614a5-aed3-4942-9382-2981e22773ec" containerName="registry-server" Jan 29 16:24:08 crc kubenswrapper[4714]: I0129 16:24:08.558509 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp" Jan 29 16:24:08 crc kubenswrapper[4714]: I0129 16:24:08.560850 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp"] Jan 29 16:24:08 crc kubenswrapper[4714]: I0129 16:24:08.561652 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-hwqbr" Jan 29 16:24:08 crc kubenswrapper[4714]: I0129 16:24:08.663567 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cm9m\" (UniqueName: \"kubernetes.io/projected/949d7185-7f54-44dd-9da9-3ed2c3c80e31-kube-api-access-4cm9m\") pod \"f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp\" (UID: \"949d7185-7f54-44dd-9da9-3ed2c3c80e31\") " pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp" Jan 29 16:24:08 crc kubenswrapper[4714]: I0129 16:24:08.663678 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/949d7185-7f54-44dd-9da9-3ed2c3c80e31-util\") pod \"f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp\" (UID: \"949d7185-7f54-44dd-9da9-3ed2c3c80e31\") " pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp" Jan 29 16:24:08 crc kubenswrapper[4714]: I0129 16:24:08.663745 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/949d7185-7f54-44dd-9da9-3ed2c3c80e31-bundle\") pod \"f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp\" (UID: \"949d7185-7f54-44dd-9da9-3ed2c3c80e31\") " pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp" Jan 29 16:24:08 crc kubenswrapper[4714]: I0129 16:24:08.764957 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/949d7185-7f54-44dd-9da9-3ed2c3c80e31-bundle\") pod \"f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp\" (UID: \"949d7185-7f54-44dd-9da9-3ed2c3c80e31\") " pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp" Jan 29 16:24:08 crc kubenswrapper[4714]: I0129 16:24:08.765012 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cm9m\" (UniqueName: \"kubernetes.io/projected/949d7185-7f54-44dd-9da9-3ed2c3c80e31-kube-api-access-4cm9m\") pod \"f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp\" (UID: \"949d7185-7f54-44dd-9da9-3ed2c3c80e31\") " pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp" Jan 29 16:24:08 crc kubenswrapper[4714]: I0129 16:24:08.765060 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/949d7185-7f54-44dd-9da9-3ed2c3c80e31-util\") pod \"f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp\" (UID: \"949d7185-7f54-44dd-9da9-3ed2c3c80e31\") " pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp" Jan 29 16:24:08 crc kubenswrapper[4714]: I0129 16:24:08.765429 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/949d7185-7f54-44dd-9da9-3ed2c3c80e31-util\") pod \"f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp\" (UID: \"949d7185-7f54-44dd-9da9-3ed2c3c80e31\") " pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp" Jan 29 16:24:08 crc kubenswrapper[4714]: I0129 16:24:08.765723 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/949d7185-7f54-44dd-9da9-3ed2c3c80e31-bundle\") pod \"f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp\" (UID: \"949d7185-7f54-44dd-9da9-3ed2c3c80e31\") " pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp" Jan 29 16:24:08 crc kubenswrapper[4714]: I0129 16:24:08.795067 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cm9m\" (UniqueName: \"kubernetes.io/projected/949d7185-7f54-44dd-9da9-3ed2c3c80e31-kube-api-access-4cm9m\") pod \"f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp\" (UID: \"949d7185-7f54-44dd-9da9-3ed2c3c80e31\") " pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp" Jan 29 16:24:08 crc kubenswrapper[4714]: I0129 16:24:08.935876 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp" Jan 29 16:24:09 crc kubenswrapper[4714]: I0129 16:24:09.363135 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp"] Jan 29 16:24:09 crc kubenswrapper[4714]: I0129 16:24:09.969044 4714 generic.go:334] "Generic (PLEG): container finished" podID="949d7185-7f54-44dd-9da9-3ed2c3c80e31" containerID="bfc5719dfa5a492d30c2a6be943eb0655e0af9b3224bd4745d65e5929dc3407a" exitCode=0 Jan 29 16:24:09 crc kubenswrapper[4714]: I0129 16:24:09.969086 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp" event={"ID":"949d7185-7f54-44dd-9da9-3ed2c3c80e31","Type":"ContainerDied","Data":"bfc5719dfa5a492d30c2a6be943eb0655e0af9b3224bd4745d65e5929dc3407a"} Jan 29 16:24:09 crc kubenswrapper[4714]: I0129 16:24:09.969111 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp" event={"ID":"949d7185-7f54-44dd-9da9-3ed2c3c80e31","Type":"ContainerStarted","Data":"62108aeab22c92a3ca54960feac1e2ad547900c8771f3b5ba2b6aee8e4a745b6"} Jan 29 16:24:10 crc kubenswrapper[4714]: I0129 16:24:10.977790 4714 generic.go:334] "Generic (PLEG): container finished" podID="949d7185-7f54-44dd-9da9-3ed2c3c80e31" containerID="6ee4a0f3d055cfa18b2c55afd177524902b4dc64f544a61af9b1f46505e17336" exitCode=0 Jan 29 16:24:10 crc kubenswrapper[4714]: I0129 16:24:10.977917 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp" event={"ID":"949d7185-7f54-44dd-9da9-3ed2c3c80e31","Type":"ContainerDied","Data":"6ee4a0f3d055cfa18b2c55afd177524902b4dc64f544a61af9b1f46505e17336"} Jan 29 16:24:11 crc kubenswrapper[4714]: I0129 16:24:11.987420 4714 generic.go:334] "Generic (PLEG): container finished" podID="949d7185-7f54-44dd-9da9-3ed2c3c80e31" containerID="a2afc6d59d0b69e82adfd2f0ef885392ef80db9eed2c4295483123972e972c1a" exitCode=0 Jan 29 16:24:11 crc kubenswrapper[4714]: I0129 16:24:11.987475 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp" event={"ID":"949d7185-7f54-44dd-9da9-3ed2c3c80e31","Type":"ContainerDied","Data":"a2afc6d59d0b69e82adfd2f0ef885392ef80db9eed2c4295483123972e972c1a"} Jan 29 16:24:13 crc kubenswrapper[4714]: I0129 16:24:13.284159 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp" Jan 29 16:24:13 crc kubenswrapper[4714]: I0129 16:24:13.427384 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/949d7185-7f54-44dd-9da9-3ed2c3c80e31-bundle\") pod \"949d7185-7f54-44dd-9da9-3ed2c3c80e31\" (UID: \"949d7185-7f54-44dd-9da9-3ed2c3c80e31\") " Jan 29 16:24:13 crc kubenswrapper[4714]: I0129 16:24:13.427556 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/949d7185-7f54-44dd-9da9-3ed2c3c80e31-util\") pod \"949d7185-7f54-44dd-9da9-3ed2c3c80e31\" (UID: \"949d7185-7f54-44dd-9da9-3ed2c3c80e31\") " Jan 29 16:24:13 crc kubenswrapper[4714]: I0129 16:24:13.427649 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cm9m\" (UniqueName: \"kubernetes.io/projected/949d7185-7f54-44dd-9da9-3ed2c3c80e31-kube-api-access-4cm9m\") pod \"949d7185-7f54-44dd-9da9-3ed2c3c80e31\" (UID: \"949d7185-7f54-44dd-9da9-3ed2c3c80e31\") " Jan 29 16:24:13 crc kubenswrapper[4714]: I0129 16:24:13.429327 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/949d7185-7f54-44dd-9da9-3ed2c3c80e31-bundle" (OuterVolumeSpecName: "bundle") pod "949d7185-7f54-44dd-9da9-3ed2c3c80e31" (UID: "949d7185-7f54-44dd-9da9-3ed2c3c80e31"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:24:13 crc kubenswrapper[4714]: I0129 16:24:13.433164 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/949d7185-7f54-44dd-9da9-3ed2c3c80e31-kube-api-access-4cm9m" (OuterVolumeSpecName: "kube-api-access-4cm9m") pod "949d7185-7f54-44dd-9da9-3ed2c3c80e31" (UID: "949d7185-7f54-44dd-9da9-3ed2c3c80e31"). InnerVolumeSpecName "kube-api-access-4cm9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:24:13 crc kubenswrapper[4714]: I0129 16:24:13.441899 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/949d7185-7f54-44dd-9da9-3ed2c3c80e31-util" (OuterVolumeSpecName: "util") pod "949d7185-7f54-44dd-9da9-3ed2c3c80e31" (UID: "949d7185-7f54-44dd-9da9-3ed2c3c80e31"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:24:13 crc kubenswrapper[4714]: I0129 16:24:13.529061 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cm9m\" (UniqueName: \"kubernetes.io/projected/949d7185-7f54-44dd-9da9-3ed2c3c80e31-kube-api-access-4cm9m\") on node \"crc\" DevicePath \"\"" Jan 29 16:24:13 crc kubenswrapper[4714]: I0129 16:24:13.529093 4714 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/949d7185-7f54-44dd-9da9-3ed2c3c80e31-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 16:24:13 crc kubenswrapper[4714]: I0129 16:24:13.529101 4714 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/949d7185-7f54-44dd-9da9-3ed2c3c80e31-util\") on node \"crc\" DevicePath \"\"" Jan 29 16:24:14 crc kubenswrapper[4714]: I0129 16:24:14.006333 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp" event={"ID":"949d7185-7f54-44dd-9da9-3ed2c3c80e31","Type":"ContainerDied","Data":"62108aeab22c92a3ca54960feac1e2ad547900c8771f3b5ba2b6aee8e4a745b6"} Jan 29 16:24:14 crc kubenswrapper[4714]: I0129 16:24:14.006788 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62108aeab22c92a3ca54960feac1e2ad547900c8771f3b5ba2b6aee8e4a745b6" Jan 29 16:24:14 crc kubenswrapper[4714]: I0129 16:24:14.006383 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp" Jan 29 16:24:21 crc kubenswrapper[4714]: I0129 16:24:21.689275 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn"] Jan 29 16:24:21 crc kubenswrapper[4714]: E0129 16:24:21.689880 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="949d7185-7f54-44dd-9da9-3ed2c3c80e31" containerName="util" Jan 29 16:24:21 crc kubenswrapper[4714]: I0129 16:24:21.689896 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="949d7185-7f54-44dd-9da9-3ed2c3c80e31" containerName="util" Jan 29 16:24:21 crc kubenswrapper[4714]: E0129 16:24:21.689908 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="949d7185-7f54-44dd-9da9-3ed2c3c80e31" containerName="extract" Jan 29 16:24:21 crc kubenswrapper[4714]: I0129 16:24:21.689915 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="949d7185-7f54-44dd-9da9-3ed2c3c80e31" containerName="extract" Jan 29 16:24:21 crc kubenswrapper[4714]: E0129 16:24:21.689951 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="949d7185-7f54-44dd-9da9-3ed2c3c80e31" containerName="pull" Jan 29 16:24:21 crc kubenswrapper[4714]: I0129 16:24:21.689961 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="949d7185-7f54-44dd-9da9-3ed2c3c80e31" containerName="pull" Jan 29 16:24:21 crc kubenswrapper[4714]: I0129 16:24:21.690087 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="949d7185-7f54-44dd-9da9-3ed2c3c80e31" containerName="extract" Jan 29 16:24:21 crc kubenswrapper[4714]: I0129 16:24:21.690544 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" Jan 29 16:24:21 crc kubenswrapper[4714]: I0129 16:24:21.693092 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-c568f" Jan 29 16:24:21 crc kubenswrapper[4714]: I0129 16:24:21.693098 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Jan 29 16:24:21 crc kubenswrapper[4714]: I0129 16:24:21.698098 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Jan 29 16:24:21 crc kubenswrapper[4714]: I0129 16:24:21.707445 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn"] Jan 29 16:24:21 crc kubenswrapper[4714]: I0129 16:24:21.842292 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8771e447-1cf7-43f9-bfab-6c1afd7476dc-webhook-cert\") pod \"mariadb-operator-controller-manager-7cc56799bb-ddchn\" (UID: \"8771e447-1cf7-43f9-bfab-6c1afd7476dc\") " pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" Jan 29 16:24:21 crc kubenswrapper[4714]: I0129 16:24:21.842853 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx2m7\" (UniqueName: \"kubernetes.io/projected/8771e447-1cf7-43f9-bfab-6c1afd7476dc-kube-api-access-mx2m7\") pod \"mariadb-operator-controller-manager-7cc56799bb-ddchn\" (UID: \"8771e447-1cf7-43f9-bfab-6c1afd7476dc\") " pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" Jan 29 16:24:21 crc kubenswrapper[4714]: I0129 16:24:21.842926 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8771e447-1cf7-43f9-bfab-6c1afd7476dc-apiservice-cert\") pod \"mariadb-operator-controller-manager-7cc56799bb-ddchn\" (UID: \"8771e447-1cf7-43f9-bfab-6c1afd7476dc\") " pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" Jan 29 16:24:21 crc kubenswrapper[4714]: I0129 16:24:21.944064 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8771e447-1cf7-43f9-bfab-6c1afd7476dc-webhook-cert\") pod \"mariadb-operator-controller-manager-7cc56799bb-ddchn\" (UID: \"8771e447-1cf7-43f9-bfab-6c1afd7476dc\") " pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" Jan 29 16:24:21 crc kubenswrapper[4714]: I0129 16:24:21.944518 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mx2m7\" (UniqueName: \"kubernetes.io/projected/8771e447-1cf7-43f9-bfab-6c1afd7476dc-kube-api-access-mx2m7\") pod \"mariadb-operator-controller-manager-7cc56799bb-ddchn\" (UID: \"8771e447-1cf7-43f9-bfab-6c1afd7476dc\") " pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" Jan 29 16:24:21 crc kubenswrapper[4714]: I0129 16:24:21.944746 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8771e447-1cf7-43f9-bfab-6c1afd7476dc-apiservice-cert\") pod \"mariadb-operator-controller-manager-7cc56799bb-ddchn\" (UID: \"8771e447-1cf7-43f9-bfab-6c1afd7476dc\") " pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" Jan 29 16:24:21 crc kubenswrapper[4714]: I0129 16:24:21.953193 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8771e447-1cf7-43f9-bfab-6c1afd7476dc-apiservice-cert\") pod \"mariadb-operator-controller-manager-7cc56799bb-ddchn\" (UID: \"8771e447-1cf7-43f9-bfab-6c1afd7476dc\") " pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" Jan 29 16:24:21 crc kubenswrapper[4714]: I0129 16:24:21.955462 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8771e447-1cf7-43f9-bfab-6c1afd7476dc-webhook-cert\") pod \"mariadb-operator-controller-manager-7cc56799bb-ddchn\" (UID: \"8771e447-1cf7-43f9-bfab-6c1afd7476dc\") " pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" Jan 29 16:24:21 crc kubenswrapper[4714]: I0129 16:24:21.962333 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx2m7\" (UniqueName: \"kubernetes.io/projected/8771e447-1cf7-43f9-bfab-6c1afd7476dc-kube-api-access-mx2m7\") pod \"mariadb-operator-controller-manager-7cc56799bb-ddchn\" (UID: \"8771e447-1cf7-43f9-bfab-6c1afd7476dc\") " pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" Jan 29 16:24:22 crc kubenswrapper[4714]: I0129 16:24:22.009418 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" Jan 29 16:24:22 crc kubenswrapper[4714]: I0129 16:24:22.275812 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn"] Jan 29 16:24:22 crc kubenswrapper[4714]: W0129 16:24:22.281932 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8771e447_1cf7_43f9_bfab_6c1afd7476dc.slice/crio-4ead9770f8d819bd3b2a9d514caf8dd18e92463bd9f126ba95d1cbb0e58fb71f WatchSource:0}: Error finding container 4ead9770f8d819bd3b2a9d514caf8dd18e92463bd9f126ba95d1cbb0e58fb71f: Status 404 returned error can't find the container with id 4ead9770f8d819bd3b2a9d514caf8dd18e92463bd9f126ba95d1cbb0e58fb71f Jan 29 16:24:23 crc kubenswrapper[4714]: I0129 16:24:23.062374 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" event={"ID":"8771e447-1cf7-43f9-bfab-6c1afd7476dc","Type":"ContainerStarted","Data":"4ead9770f8d819bd3b2a9d514caf8dd18e92463bd9f126ba95d1cbb0e58fb71f"} Jan 29 16:24:26 crc kubenswrapper[4714]: I0129 16:24:26.084107 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" event={"ID":"8771e447-1cf7-43f9-bfab-6c1afd7476dc","Type":"ContainerStarted","Data":"265cc03bd32fdd618d4ab75713d4df2f554d8e322232952f23e67ae7895f0208"} Jan 29 16:24:26 crc kubenswrapper[4714]: I0129 16:24:26.084696 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" Jan 29 16:24:26 crc kubenswrapper[4714]: I0129 16:24:26.102950 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" podStartSLOduration=1.861245273 podStartE2EDuration="5.102921372s" podCreationTimestamp="2026-01-29 16:24:21 +0000 UTC" firstStartedPulling="2026-01-29 16:24:22.284660459 +0000 UTC m=+868.805161589" lastFinishedPulling="2026-01-29 16:24:25.526336578 +0000 UTC m=+872.046837688" observedRunningTime="2026-01-29 16:24:26.101012202 +0000 UTC m=+872.621513332" watchObservedRunningTime="2026-01-29 16:24:26.102921372 +0000 UTC m=+872.623422492" Jan 29 16:24:32 crc kubenswrapper[4714]: I0129 16:24:32.015388 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" Jan 29 16:24:38 crc kubenswrapper[4714]: I0129 16:24:38.449955 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-44nbs"] Jan 29 16:24:38 crc kubenswrapper[4714]: I0129 16:24:38.451402 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-44nbs" Jan 29 16:24:38 crc kubenswrapper[4714]: I0129 16:24:38.454444 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-5fnhf" Jan 29 16:24:38 crc kubenswrapper[4714]: I0129 16:24:38.457989 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-44nbs"] Jan 29 16:24:38 crc kubenswrapper[4714]: I0129 16:24:38.486854 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhzpl\" (UniqueName: \"kubernetes.io/projected/6e7fe80f-e880-4c2d-8f2e-1861fb1575b7-kube-api-access-zhzpl\") pod \"infra-operator-index-44nbs\" (UID: \"6e7fe80f-e880-4c2d-8f2e-1861fb1575b7\") " pod="openstack-operators/infra-operator-index-44nbs" Jan 29 16:24:38 crc kubenswrapper[4714]: I0129 16:24:38.588027 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhzpl\" (UniqueName: \"kubernetes.io/projected/6e7fe80f-e880-4c2d-8f2e-1861fb1575b7-kube-api-access-zhzpl\") pod \"infra-operator-index-44nbs\" (UID: \"6e7fe80f-e880-4c2d-8f2e-1861fb1575b7\") " pod="openstack-operators/infra-operator-index-44nbs" Jan 29 16:24:38 crc kubenswrapper[4714]: I0129 16:24:38.626895 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhzpl\" (UniqueName: \"kubernetes.io/projected/6e7fe80f-e880-4c2d-8f2e-1861fb1575b7-kube-api-access-zhzpl\") pod \"infra-operator-index-44nbs\" (UID: \"6e7fe80f-e880-4c2d-8f2e-1861fb1575b7\") " pod="openstack-operators/infra-operator-index-44nbs" Jan 29 16:24:38 crc kubenswrapper[4714]: I0129 16:24:38.771247 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-44nbs" Jan 29 16:24:40 crc kubenswrapper[4714]: I0129 16:24:40.025258 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-44nbs"] Jan 29 16:24:40 crc kubenswrapper[4714]: I0129 16:24:40.166498 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-44nbs" event={"ID":"6e7fe80f-e880-4c2d-8f2e-1861fb1575b7","Type":"ContainerStarted","Data":"2b032bfed2b84d9c544866a2e2b9f0d7688dc192688bc1e69bbc65eeaf0fbe29"} Jan 29 16:24:42 crc kubenswrapper[4714]: I0129 16:24:42.196611 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-44nbs" event={"ID":"6e7fe80f-e880-4c2d-8f2e-1861fb1575b7","Type":"ContainerStarted","Data":"3da4379d19ac2f9c90609a5bbb56d1adb9a0376fa03d524802535beb518bf498"} Jan 29 16:24:42 crc kubenswrapper[4714]: I0129 16:24:42.214147 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-44nbs" podStartSLOduration=2.906302288 podStartE2EDuration="4.214116376s" podCreationTimestamp="2026-01-29 16:24:38 +0000 UTC" firstStartedPulling="2026-01-29 16:24:40.043260627 +0000 UTC m=+886.563761747" lastFinishedPulling="2026-01-29 16:24:41.351074695 +0000 UTC m=+887.871575835" observedRunningTime="2026-01-29 16:24:42.208181851 +0000 UTC m=+888.728683001" watchObservedRunningTime="2026-01-29 16:24:42.214116376 +0000 UTC m=+888.734617536" Jan 29 16:24:42 crc kubenswrapper[4714]: I0129 16:24:42.442773 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-44nbs"] Jan 29 16:24:43 crc kubenswrapper[4714]: I0129 16:24:43.052650 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-9xq82"] Jan 29 16:24:43 crc kubenswrapper[4714]: I0129 16:24:43.054376 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-9xq82" Jan 29 16:24:43 crc kubenswrapper[4714]: I0129 16:24:43.057858 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-9xq82"] Jan 29 16:24:43 crc kubenswrapper[4714]: I0129 16:24:43.159498 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ts4kr\" (UniqueName: \"kubernetes.io/projected/9dcd8561-aa17-46a8-b184-0495c320a33b-kube-api-access-ts4kr\") pod \"infra-operator-index-9xq82\" (UID: \"9dcd8561-aa17-46a8-b184-0495c320a33b\") " pod="openstack-operators/infra-operator-index-9xq82" Jan 29 16:24:43 crc kubenswrapper[4714]: I0129 16:24:43.260873 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ts4kr\" (UniqueName: \"kubernetes.io/projected/9dcd8561-aa17-46a8-b184-0495c320a33b-kube-api-access-ts4kr\") pod \"infra-operator-index-9xq82\" (UID: \"9dcd8561-aa17-46a8-b184-0495c320a33b\") " pod="openstack-operators/infra-operator-index-9xq82" Jan 29 16:24:43 crc kubenswrapper[4714]: I0129 16:24:43.283770 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ts4kr\" (UniqueName: \"kubernetes.io/projected/9dcd8561-aa17-46a8-b184-0495c320a33b-kube-api-access-ts4kr\") pod \"infra-operator-index-9xq82\" (UID: \"9dcd8561-aa17-46a8-b184-0495c320a33b\") " pod="openstack-operators/infra-operator-index-9xq82" Jan 29 16:24:43 crc kubenswrapper[4714]: I0129 16:24:43.375753 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-9xq82" Jan 29 16:24:43 crc kubenswrapper[4714]: I0129 16:24:43.564804 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-9xq82"] Jan 29 16:24:44 crc kubenswrapper[4714]: I0129 16:24:44.199637 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-9xq82" event={"ID":"9dcd8561-aa17-46a8-b184-0495c320a33b","Type":"ContainerStarted","Data":"17fa7e5da52af363967b26a09672574689ebcabaf095f6c1443af9a04ea81254"} Jan 29 16:24:44 crc kubenswrapper[4714]: I0129 16:24:44.199704 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-9xq82" event={"ID":"9dcd8561-aa17-46a8-b184-0495c320a33b","Type":"ContainerStarted","Data":"b25ca92cca2c102702aabf44744c607b11ac82433ac2ea8b9c60134c6952d1ea"} Jan 29 16:24:44 crc kubenswrapper[4714]: I0129 16:24:44.199758 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-44nbs" podUID="6e7fe80f-e880-4c2d-8f2e-1861fb1575b7" containerName="registry-server" containerID="cri-o://3da4379d19ac2f9c90609a5bbb56d1adb9a0376fa03d524802535beb518bf498" gracePeriod=2 Jan 29 16:24:44 crc kubenswrapper[4714]: I0129 16:24:44.229627 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-9xq82" podStartSLOduration=0.786167259 podStartE2EDuration="1.229601073s" podCreationTimestamp="2026-01-29 16:24:43 +0000 UTC" firstStartedPulling="2026-01-29 16:24:43.574020773 +0000 UTC m=+890.094521893" lastFinishedPulling="2026-01-29 16:24:44.017454587 +0000 UTC m=+890.537955707" observedRunningTime="2026-01-29 16:24:44.226907569 +0000 UTC m=+890.747408709" watchObservedRunningTime="2026-01-29 16:24:44.229601073 +0000 UTC m=+890.750102193" Jan 29 16:24:44 crc kubenswrapper[4714]: I0129 16:24:44.520404 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-44nbs" Jan 29 16:24:44 crc kubenswrapper[4714]: I0129 16:24:44.592323 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhzpl\" (UniqueName: \"kubernetes.io/projected/6e7fe80f-e880-4c2d-8f2e-1861fb1575b7-kube-api-access-zhzpl\") pod \"6e7fe80f-e880-4c2d-8f2e-1861fb1575b7\" (UID: \"6e7fe80f-e880-4c2d-8f2e-1861fb1575b7\") " Jan 29 16:24:44 crc kubenswrapper[4714]: I0129 16:24:44.599237 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e7fe80f-e880-4c2d-8f2e-1861fb1575b7-kube-api-access-zhzpl" (OuterVolumeSpecName: "kube-api-access-zhzpl") pod "6e7fe80f-e880-4c2d-8f2e-1861fb1575b7" (UID: "6e7fe80f-e880-4c2d-8f2e-1861fb1575b7"). InnerVolumeSpecName "kube-api-access-zhzpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:24:44 crc kubenswrapper[4714]: I0129 16:24:44.693702 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhzpl\" (UniqueName: \"kubernetes.io/projected/6e7fe80f-e880-4c2d-8f2e-1861fb1575b7-kube-api-access-zhzpl\") on node \"crc\" DevicePath \"\"" Jan 29 16:24:45 crc kubenswrapper[4714]: I0129 16:24:45.207723 4714 generic.go:334] "Generic (PLEG): container finished" podID="6e7fe80f-e880-4c2d-8f2e-1861fb1575b7" containerID="3da4379d19ac2f9c90609a5bbb56d1adb9a0376fa03d524802535beb518bf498" exitCode=0 Jan 29 16:24:45 crc kubenswrapper[4714]: I0129 16:24:45.207791 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-44nbs" Jan 29 16:24:45 crc kubenswrapper[4714]: I0129 16:24:45.207794 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-44nbs" event={"ID":"6e7fe80f-e880-4c2d-8f2e-1861fb1575b7","Type":"ContainerDied","Data":"3da4379d19ac2f9c90609a5bbb56d1adb9a0376fa03d524802535beb518bf498"} Jan 29 16:24:45 crc kubenswrapper[4714]: I0129 16:24:45.207890 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-44nbs" event={"ID":"6e7fe80f-e880-4c2d-8f2e-1861fb1575b7","Type":"ContainerDied","Data":"2b032bfed2b84d9c544866a2e2b9f0d7688dc192688bc1e69bbc65eeaf0fbe29"} Jan 29 16:24:45 crc kubenswrapper[4714]: I0129 16:24:45.207965 4714 scope.go:117] "RemoveContainer" containerID="3da4379d19ac2f9c90609a5bbb56d1adb9a0376fa03d524802535beb518bf498" Jan 29 16:24:45 crc kubenswrapper[4714]: I0129 16:24:45.232741 4714 scope.go:117] "RemoveContainer" containerID="3da4379d19ac2f9c90609a5bbb56d1adb9a0376fa03d524802535beb518bf498" Jan 29 16:24:45 crc kubenswrapper[4714]: E0129 16:24:45.233954 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3da4379d19ac2f9c90609a5bbb56d1adb9a0376fa03d524802535beb518bf498\": container with ID starting with 3da4379d19ac2f9c90609a5bbb56d1adb9a0376fa03d524802535beb518bf498 not found: ID does not exist" containerID="3da4379d19ac2f9c90609a5bbb56d1adb9a0376fa03d524802535beb518bf498" Jan 29 16:24:45 crc kubenswrapper[4714]: I0129 16:24:45.234000 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3da4379d19ac2f9c90609a5bbb56d1adb9a0376fa03d524802535beb518bf498"} err="failed to get container status \"3da4379d19ac2f9c90609a5bbb56d1adb9a0376fa03d524802535beb518bf498\": rpc error: code = NotFound desc = could not find container \"3da4379d19ac2f9c90609a5bbb56d1adb9a0376fa03d524802535beb518bf498\": container with ID starting with 3da4379d19ac2f9c90609a5bbb56d1adb9a0376fa03d524802535beb518bf498 not found: ID does not exist" Jan 29 16:24:45 crc kubenswrapper[4714]: I0129 16:24:45.256544 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-44nbs"] Jan 29 16:24:45 crc kubenswrapper[4714]: I0129 16:24:45.261639 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-44nbs"] Jan 29 16:24:46 crc kubenswrapper[4714]: I0129 16:24:46.194505 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e7fe80f-e880-4c2d-8f2e-1861fb1575b7" path="/var/lib/kubelet/pods/6e7fe80f-e880-4c2d-8f2e-1861fb1575b7/volumes" Jan 29 16:24:53 crc kubenswrapper[4714]: I0129 16:24:53.376725 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-9xq82" Jan 29 16:24:53 crc kubenswrapper[4714]: I0129 16:24:53.377604 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-9xq82" Jan 29 16:24:53 crc kubenswrapper[4714]: I0129 16:24:53.411626 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-9xq82" Jan 29 16:24:54 crc kubenswrapper[4714]: I0129 16:24:54.299068 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-9xq82" Jan 29 16:24:56 crc kubenswrapper[4714]: I0129 16:24:56.301356 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs"] Jan 29 16:24:56 crc kubenswrapper[4714]: E0129 16:24:56.301996 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e7fe80f-e880-4c2d-8f2e-1861fb1575b7" containerName="registry-server" Jan 29 16:24:56 crc kubenswrapper[4714]: I0129 16:24:56.302012 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e7fe80f-e880-4c2d-8f2e-1861fb1575b7" containerName="registry-server" Jan 29 16:24:56 crc kubenswrapper[4714]: I0129 16:24:56.302139 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e7fe80f-e880-4c2d-8f2e-1861fb1575b7" containerName="registry-server" Jan 29 16:24:56 crc kubenswrapper[4714]: I0129 16:24:56.303042 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs" Jan 29 16:24:56 crc kubenswrapper[4714]: I0129 16:24:56.306045 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-hwqbr" Jan 29 16:24:56 crc kubenswrapper[4714]: I0129 16:24:56.309329 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs"] Jan 29 16:24:56 crc kubenswrapper[4714]: I0129 16:24:56.452870 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/096bd691-cca6-4566-b56c-7643e2feaef1-bundle\") pod \"d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs\" (UID: \"096bd691-cca6-4566-b56c-7643e2feaef1\") " pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs" Jan 29 16:24:56 crc kubenswrapper[4714]: I0129 16:24:56.453141 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pplc7\" (UniqueName: \"kubernetes.io/projected/096bd691-cca6-4566-b56c-7643e2feaef1-kube-api-access-pplc7\") pod \"d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs\" (UID: \"096bd691-cca6-4566-b56c-7643e2feaef1\") " pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs" Jan 29 16:24:56 crc kubenswrapper[4714]: I0129 16:24:56.453222 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/096bd691-cca6-4566-b56c-7643e2feaef1-util\") pod \"d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs\" (UID: \"096bd691-cca6-4566-b56c-7643e2feaef1\") " pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs" Jan 29 16:24:56 crc kubenswrapper[4714]: I0129 16:24:56.554619 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/096bd691-cca6-4566-b56c-7643e2feaef1-bundle\") pod \"d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs\" (UID: \"096bd691-cca6-4566-b56c-7643e2feaef1\") " pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs" Jan 29 16:24:56 crc kubenswrapper[4714]: I0129 16:24:56.554810 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pplc7\" (UniqueName: \"kubernetes.io/projected/096bd691-cca6-4566-b56c-7643e2feaef1-kube-api-access-pplc7\") pod \"d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs\" (UID: \"096bd691-cca6-4566-b56c-7643e2feaef1\") " pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs" Jan 29 16:24:56 crc kubenswrapper[4714]: I0129 16:24:56.554880 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/096bd691-cca6-4566-b56c-7643e2feaef1-util\") pod \"d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs\" (UID: \"096bd691-cca6-4566-b56c-7643e2feaef1\") " pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs" Jan 29 16:24:56 crc kubenswrapper[4714]: I0129 16:24:56.555133 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/096bd691-cca6-4566-b56c-7643e2feaef1-bundle\") pod \"d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs\" (UID: \"096bd691-cca6-4566-b56c-7643e2feaef1\") " pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs" Jan 29 16:24:56 crc kubenswrapper[4714]: I0129 16:24:56.555161 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/096bd691-cca6-4566-b56c-7643e2feaef1-util\") pod \"d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs\" (UID: \"096bd691-cca6-4566-b56c-7643e2feaef1\") " pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs" Jan 29 16:24:56 crc kubenswrapper[4714]: I0129 16:24:56.573981 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pplc7\" (UniqueName: \"kubernetes.io/projected/096bd691-cca6-4566-b56c-7643e2feaef1-kube-api-access-pplc7\") pod \"d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs\" (UID: \"096bd691-cca6-4566-b56c-7643e2feaef1\") " pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs" Jan 29 16:24:56 crc kubenswrapper[4714]: I0129 16:24:56.621778 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs" Jan 29 16:24:57 crc kubenswrapper[4714]: I0129 16:24:57.054783 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs"] Jan 29 16:24:57 crc kubenswrapper[4714]: I0129 16:24:57.283081 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs" event={"ID":"096bd691-cca6-4566-b56c-7643e2feaef1","Type":"ContainerStarted","Data":"1cd564e322fa134ca0b1e14f7b1c05c0de31a1e7c8f443cd6b64bbf340b9a6ae"} Jan 29 16:24:57 crc kubenswrapper[4714]: I0129 16:24:57.283130 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs" event={"ID":"096bd691-cca6-4566-b56c-7643e2feaef1","Type":"ContainerStarted","Data":"9498d20df0cd78103e7153f16a6de5efe6b72b7790741a99a88e73ae8cbf811d"} Jan 29 16:24:58 crc kubenswrapper[4714]: I0129 16:24:58.291654 4714 generic.go:334] "Generic (PLEG): container finished" podID="096bd691-cca6-4566-b56c-7643e2feaef1" containerID="1cd564e322fa134ca0b1e14f7b1c05c0de31a1e7c8f443cd6b64bbf340b9a6ae" exitCode=0 Jan 29 16:24:58 crc kubenswrapper[4714]: I0129 16:24:58.291711 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs" event={"ID":"096bd691-cca6-4566-b56c-7643e2feaef1","Type":"ContainerDied","Data":"1cd564e322fa134ca0b1e14f7b1c05c0de31a1e7c8f443cd6b64bbf340b9a6ae"} Jan 29 16:25:00 crc kubenswrapper[4714]: I0129 16:25:00.312596 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs" event={"ID":"096bd691-cca6-4566-b56c-7643e2feaef1","Type":"ContainerStarted","Data":"8be88780f2ccaa67529f7c97f45a315a79167378ebf7e1fdcde84818ec246373"} Jan 29 16:25:01 crc kubenswrapper[4714]: I0129 16:25:01.322047 4714 generic.go:334] "Generic (PLEG): container finished" podID="096bd691-cca6-4566-b56c-7643e2feaef1" containerID="8be88780f2ccaa67529f7c97f45a315a79167378ebf7e1fdcde84818ec246373" exitCode=0 Jan 29 16:25:01 crc kubenswrapper[4714]: I0129 16:25:01.322099 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs" event={"ID":"096bd691-cca6-4566-b56c-7643e2feaef1","Type":"ContainerDied","Data":"8be88780f2ccaa67529f7c97f45a315a79167378ebf7e1fdcde84818ec246373"} Jan 29 16:25:02 crc kubenswrapper[4714]: I0129 16:25:02.333544 4714 generic.go:334] "Generic (PLEG): container finished" podID="096bd691-cca6-4566-b56c-7643e2feaef1" containerID="3bc2a2089f6be296701e5d57c79cbb2a9a2dd560e4db7f4ea6460bad3386ed41" exitCode=0 Jan 29 16:25:02 crc kubenswrapper[4714]: I0129 16:25:02.333666 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs" event={"ID":"096bd691-cca6-4566-b56c-7643e2feaef1","Type":"ContainerDied","Data":"3bc2a2089f6be296701e5d57c79cbb2a9a2dd560e4db7f4ea6460bad3386ed41"} Jan 29 16:25:03 crc kubenswrapper[4714]: I0129 16:25:03.588520 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs" Jan 29 16:25:03 crc kubenswrapper[4714]: I0129 16:25:03.754642 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/096bd691-cca6-4566-b56c-7643e2feaef1-util\") pod \"096bd691-cca6-4566-b56c-7643e2feaef1\" (UID: \"096bd691-cca6-4566-b56c-7643e2feaef1\") " Jan 29 16:25:03 crc kubenswrapper[4714]: I0129 16:25:03.754745 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pplc7\" (UniqueName: \"kubernetes.io/projected/096bd691-cca6-4566-b56c-7643e2feaef1-kube-api-access-pplc7\") pod \"096bd691-cca6-4566-b56c-7643e2feaef1\" (UID: \"096bd691-cca6-4566-b56c-7643e2feaef1\") " Jan 29 16:25:03 crc kubenswrapper[4714]: I0129 16:25:03.754783 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/096bd691-cca6-4566-b56c-7643e2feaef1-bundle\") pod \"096bd691-cca6-4566-b56c-7643e2feaef1\" (UID: \"096bd691-cca6-4566-b56c-7643e2feaef1\") " Jan 29 16:25:03 crc kubenswrapper[4714]: I0129 16:25:03.757272 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/096bd691-cca6-4566-b56c-7643e2feaef1-bundle" (OuterVolumeSpecName: "bundle") pod "096bd691-cca6-4566-b56c-7643e2feaef1" (UID: "096bd691-cca6-4566-b56c-7643e2feaef1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:25:03 crc kubenswrapper[4714]: I0129 16:25:03.765134 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/096bd691-cca6-4566-b56c-7643e2feaef1-kube-api-access-pplc7" (OuterVolumeSpecName: "kube-api-access-pplc7") pod "096bd691-cca6-4566-b56c-7643e2feaef1" (UID: "096bd691-cca6-4566-b56c-7643e2feaef1"). InnerVolumeSpecName "kube-api-access-pplc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:25:03 crc kubenswrapper[4714]: I0129 16:25:03.765755 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/096bd691-cca6-4566-b56c-7643e2feaef1-util" (OuterVolumeSpecName: "util") pod "096bd691-cca6-4566-b56c-7643e2feaef1" (UID: "096bd691-cca6-4566-b56c-7643e2feaef1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:25:03 crc kubenswrapper[4714]: I0129 16:25:03.855382 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pplc7\" (UniqueName: \"kubernetes.io/projected/096bd691-cca6-4566-b56c-7643e2feaef1-kube-api-access-pplc7\") on node \"crc\" DevicePath \"\"" Jan 29 16:25:03 crc kubenswrapper[4714]: I0129 16:25:03.855424 4714 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/096bd691-cca6-4566-b56c-7643e2feaef1-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 16:25:03 crc kubenswrapper[4714]: I0129 16:25:03.855432 4714 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/096bd691-cca6-4566-b56c-7643e2feaef1-util\") on node \"crc\" DevicePath \"\"" Jan 29 16:25:04 crc kubenswrapper[4714]: I0129 16:25:04.348133 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs" event={"ID":"096bd691-cca6-4566-b56c-7643e2feaef1","Type":"ContainerDied","Data":"9498d20df0cd78103e7153f16a6de5efe6b72b7790741a99a88e73ae8cbf811d"} Jan 29 16:25:04 crc kubenswrapper[4714]: I0129 16:25:04.348206 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9498d20df0cd78103e7153f16a6de5efe6b72b7790741a99a88e73ae8cbf811d" Jan 29 16:25:04 crc kubenswrapper[4714]: I0129 16:25:04.348230 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs" Jan 29 16:25:13 crc kubenswrapper[4714]: I0129 16:25:13.529643 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq"] Jan 29 16:25:13 crc kubenswrapper[4714]: E0129 16:25:13.530228 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="096bd691-cca6-4566-b56c-7643e2feaef1" containerName="pull" Jan 29 16:25:13 crc kubenswrapper[4714]: I0129 16:25:13.530265 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="096bd691-cca6-4566-b56c-7643e2feaef1" containerName="pull" Jan 29 16:25:13 crc kubenswrapper[4714]: E0129 16:25:13.530282 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="096bd691-cca6-4566-b56c-7643e2feaef1" containerName="extract" Jan 29 16:25:13 crc kubenswrapper[4714]: I0129 16:25:13.530290 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="096bd691-cca6-4566-b56c-7643e2feaef1" containerName="extract" Jan 29 16:25:13 crc kubenswrapper[4714]: E0129 16:25:13.530309 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="096bd691-cca6-4566-b56c-7643e2feaef1" containerName="util" Jan 29 16:25:13 crc kubenswrapper[4714]: I0129 16:25:13.530317 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="096bd691-cca6-4566-b56c-7643e2feaef1" containerName="util" Jan 29 16:25:13 crc kubenswrapper[4714]: I0129 16:25:13.530474 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="096bd691-cca6-4566-b56c-7643e2feaef1" containerName="extract" Jan 29 16:25:13 crc kubenswrapper[4714]: I0129 16:25:13.530960 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" Jan 29 16:25:13 crc kubenswrapper[4714]: I0129 16:25:13.532906 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-5tqg8" Jan 29 16:25:13 crc kubenswrapper[4714]: I0129 16:25:13.533301 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Jan 29 16:25:13 crc kubenswrapper[4714]: I0129 16:25:13.543835 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq"] Jan 29 16:25:13 crc kubenswrapper[4714]: I0129 16:25:13.689148 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2779e724-225f-4a5f-9e2c-3b05fe08dff2-apiservice-cert\") pod \"infra-operator-controller-manager-66f4f5476c-xqnxq\" (UID: \"2779e724-225f-4a5f-9e2c-3b05fe08dff2\") " pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" Jan 29 16:25:13 crc kubenswrapper[4714]: I0129 16:25:13.689217 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2779e724-225f-4a5f-9e2c-3b05fe08dff2-webhook-cert\") pod \"infra-operator-controller-manager-66f4f5476c-xqnxq\" (UID: \"2779e724-225f-4a5f-9e2c-3b05fe08dff2\") " pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" Jan 29 16:25:13 crc kubenswrapper[4714]: I0129 16:25:13.689286 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jxvt\" (UniqueName: \"kubernetes.io/projected/2779e724-225f-4a5f-9e2c-3b05fe08dff2-kube-api-access-6jxvt\") pod \"infra-operator-controller-manager-66f4f5476c-xqnxq\" (UID: \"2779e724-225f-4a5f-9e2c-3b05fe08dff2\") " pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" Jan 29 16:25:13 crc kubenswrapper[4714]: I0129 16:25:13.790010 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jxvt\" (UniqueName: \"kubernetes.io/projected/2779e724-225f-4a5f-9e2c-3b05fe08dff2-kube-api-access-6jxvt\") pod \"infra-operator-controller-manager-66f4f5476c-xqnxq\" (UID: \"2779e724-225f-4a5f-9e2c-3b05fe08dff2\") " pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" Jan 29 16:25:13 crc kubenswrapper[4714]: I0129 16:25:13.790105 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2779e724-225f-4a5f-9e2c-3b05fe08dff2-apiservice-cert\") pod \"infra-operator-controller-manager-66f4f5476c-xqnxq\" (UID: \"2779e724-225f-4a5f-9e2c-3b05fe08dff2\") " pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" Jan 29 16:25:13 crc kubenswrapper[4714]: I0129 16:25:13.790138 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2779e724-225f-4a5f-9e2c-3b05fe08dff2-webhook-cert\") pod \"infra-operator-controller-manager-66f4f5476c-xqnxq\" (UID: \"2779e724-225f-4a5f-9e2c-3b05fe08dff2\") " pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" Jan 29 16:25:13 crc kubenswrapper[4714]: I0129 16:25:13.795662 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2779e724-225f-4a5f-9e2c-3b05fe08dff2-webhook-cert\") pod \"infra-operator-controller-manager-66f4f5476c-xqnxq\" (UID: \"2779e724-225f-4a5f-9e2c-3b05fe08dff2\") " pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" Jan 29 16:25:13 crc kubenswrapper[4714]: I0129 16:25:13.795695 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2779e724-225f-4a5f-9e2c-3b05fe08dff2-apiservice-cert\") pod \"infra-operator-controller-manager-66f4f5476c-xqnxq\" (UID: \"2779e724-225f-4a5f-9e2c-3b05fe08dff2\") " pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" Jan 29 16:25:13 crc kubenswrapper[4714]: I0129 16:25:13.810114 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jxvt\" (UniqueName: \"kubernetes.io/projected/2779e724-225f-4a5f-9e2c-3b05fe08dff2-kube-api-access-6jxvt\") pod \"infra-operator-controller-manager-66f4f5476c-xqnxq\" (UID: \"2779e724-225f-4a5f-9e2c-3b05fe08dff2\") " pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" Jan 29 16:25:13 crc kubenswrapper[4714]: I0129 16:25:13.868470 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" Jan 29 16:25:14 crc kubenswrapper[4714]: I0129 16:25:14.274803 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq"] Jan 29 16:25:14 crc kubenswrapper[4714]: I0129 16:25:14.415157 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" event={"ID":"2779e724-225f-4a5f-9e2c-3b05fe08dff2","Type":"ContainerStarted","Data":"a29d14b8a75b07c5bbb8bc497bdaef3d2bf58ebfc2cc63259d7bb78d82a74639"} Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.283817 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.286489 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.291334 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"galera-openstack-dockercfg-qdqvq" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.291424 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"openstack-config-data" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.291570 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"openstack-scripts" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.292463 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"kube-root-ca.crt" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.292556 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"openshift-service-ca.crt" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.303249 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.308372 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.309444 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.314975 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.316410 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.325532 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.331334 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.447888 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.447980 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f8d336f2-b190-4e32-be3a-27fbf0e50a06-kolla-config\") pod \"openstack-galera-1\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.448021 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-2\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.448057 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e367e739-45d9-4c71-82fa-ecda02da3277-kolla-config\") pod \"openstack-galera-2\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.448083 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f8d336f2-b190-4e32-be3a-27fbf0e50a06-config-data-generated\") pod \"openstack-galera-1\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.448103 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e27f02c1-a7d5-4d49-838b-df5445720a07-kolla-config\") pod \"openstack-galera-0\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.448126 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e27f02c1-a7d5-4d49-838b-df5445720a07-config-data-default\") pod \"openstack-galera-0\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.448157 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-1\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.448301 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f8d336f2-b190-4e32-be3a-27fbf0e50a06-operator-scripts\") pod \"openstack-galera-1\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.448364 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e27f02c1-a7d5-4d49-838b-df5445720a07-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.448389 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e367e739-45d9-4c71-82fa-ecda02da3277-config-data-generated\") pod \"openstack-galera-2\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.448460 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhbgj\" (UniqueName: \"kubernetes.io/projected/f8d336f2-b190-4e32-be3a-27fbf0e50a06-kube-api-access-lhbgj\") pod \"openstack-galera-1\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.448490 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6vvb\" (UniqueName: \"kubernetes.io/projected/e27f02c1-a7d5-4d49-838b-df5445720a07-kube-api-access-p6vvb\") pod \"openstack-galera-0\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.448506 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nmlf\" (UniqueName: \"kubernetes.io/projected/e367e739-45d9-4c71-82fa-ecda02da3277-kube-api-access-8nmlf\") pod \"openstack-galera-2\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.448527 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f8d336f2-b190-4e32-be3a-27fbf0e50a06-config-data-default\") pod \"openstack-galera-1\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.448769 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e367e739-45d9-4c71-82fa-ecda02da3277-operator-scripts\") pod \"openstack-galera-2\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.448824 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e367e739-45d9-4c71-82fa-ecda02da3277-config-data-default\") pod \"openstack-galera-2\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.448860 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e27f02c1-a7d5-4d49-838b-df5445720a07-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.550274 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f8d336f2-b190-4e32-be3a-27fbf0e50a06-operator-scripts\") pod \"openstack-galera-1\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.550318 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e27f02c1-a7d5-4d49-838b-df5445720a07-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.550342 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e367e739-45d9-4c71-82fa-ecda02da3277-config-data-generated\") pod \"openstack-galera-2\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.550371 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6vvb\" (UniqueName: \"kubernetes.io/projected/e27f02c1-a7d5-4d49-838b-df5445720a07-kube-api-access-p6vvb\") pod \"openstack-galera-0\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.550389 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nmlf\" (UniqueName: \"kubernetes.io/projected/e367e739-45d9-4c71-82fa-ecda02da3277-kube-api-access-8nmlf\") pod \"openstack-galera-2\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.550407 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhbgj\" (UniqueName: \"kubernetes.io/projected/f8d336f2-b190-4e32-be3a-27fbf0e50a06-kube-api-access-lhbgj\") pod \"openstack-galera-1\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.550429 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f8d336f2-b190-4e32-be3a-27fbf0e50a06-config-data-default\") pod \"openstack-galera-1\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.550468 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e367e739-45d9-4c71-82fa-ecda02da3277-operator-scripts\") pod \"openstack-galera-2\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.550494 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e367e739-45d9-4c71-82fa-ecda02da3277-config-data-default\") pod \"openstack-galera-2\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.550516 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e27f02c1-a7d5-4d49-838b-df5445720a07-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.550541 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.550560 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f8d336f2-b190-4e32-be3a-27fbf0e50a06-kolla-config\") pod \"openstack-galera-1\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.550595 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-2\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.550620 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e367e739-45d9-4c71-82fa-ecda02da3277-kolla-config\") pod \"openstack-galera-2\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.550641 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e27f02c1-a7d5-4d49-838b-df5445720a07-kolla-config\") pod \"openstack-galera-0\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.550661 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f8d336f2-b190-4e32-be3a-27fbf0e50a06-config-data-generated\") pod \"openstack-galera-1\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.550682 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e27f02c1-a7d5-4d49-838b-df5445720a07-config-data-default\") pod \"openstack-galera-0\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.550708 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-1\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.551588 4714 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-1\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") device mount path \"/mnt/openstack/pv04\"" pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.551591 4714 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") device mount path \"/mnt/openstack/pv01\"" pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.551622 4714 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-2\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") device mount path \"/mnt/openstack/pv10\"" pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.556368 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e27f02c1-a7d5-4d49-838b-df5445720a07-kolla-config\") pod \"openstack-galera-0\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.556832 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f8d336f2-b190-4e32-be3a-27fbf0e50a06-config-data-default\") pod \"openstack-galera-1\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.556922 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f8d336f2-b190-4e32-be3a-27fbf0e50a06-kolla-config\") pod \"openstack-galera-1\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.557112 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f8d336f2-b190-4e32-be3a-27fbf0e50a06-config-data-generated\") pod \"openstack-galera-1\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.557195 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e27f02c1-a7d5-4d49-838b-df5445720a07-config-data-default\") pod \"openstack-galera-0\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.557382 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f8d336f2-b190-4e32-be3a-27fbf0e50a06-operator-scripts\") pod \"openstack-galera-1\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.557562 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e367e739-45d9-4c71-82fa-ecda02da3277-config-data-generated\") pod \"openstack-galera-2\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.558346 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e27f02c1-a7d5-4d49-838b-df5445720a07-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.559045 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e367e739-45d9-4c71-82fa-ecda02da3277-kolla-config\") pod \"openstack-galera-2\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.559648 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e27f02c1-a7d5-4d49-838b-df5445720a07-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.559795 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e367e739-45d9-4c71-82fa-ecda02da3277-operator-scripts\") pod \"openstack-galera-2\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.570845 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nmlf\" (UniqueName: \"kubernetes.io/projected/e367e739-45d9-4c71-82fa-ecda02da3277-kube-api-access-8nmlf\") pod \"openstack-galera-2\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.572065 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.575247 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-1\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.577275 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhbgj\" (UniqueName: \"kubernetes.io/projected/f8d336f2-b190-4e32-be3a-27fbf0e50a06-kube-api-access-lhbgj\") pod \"openstack-galera-1\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.578958 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e367e739-45d9-4c71-82fa-ecda02da3277-config-data-default\") pod \"openstack-galera-2\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.580717 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6vvb\" (UniqueName: \"kubernetes.io/projected/e27f02c1-a7d5-4d49-838b-df5445720a07-kube-api-access-p6vvb\") pod \"openstack-galera-0\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.583678 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-2\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.614835 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.635148 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:17 crc kubenswrapper[4714]: I0129 16:25:17.650587 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:18 crc kubenswrapper[4714]: I0129 16:25:18.059252 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Jan 29 16:25:18 crc kubenswrapper[4714]: I0129 16:25:18.103587 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Jan 29 16:25:18 crc kubenswrapper[4714]: W0129 16:25:18.107546 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8d336f2_b190_4e32_be3a_27fbf0e50a06.slice/crio-2730c5f20c68259dc37f00e3d986e43810a9d5fe85207a2c704b5656855e553b WatchSource:0}: Error finding container 2730c5f20c68259dc37f00e3d986e43810a9d5fe85207a2c704b5656855e553b: Status 404 returned error can't find the container with id 2730c5f20c68259dc37f00e3d986e43810a9d5fe85207a2c704b5656855e553b Jan 29 16:25:18 crc kubenswrapper[4714]: I0129 16:25:18.111496 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Jan 29 16:25:18 crc kubenswrapper[4714]: W0129 16:25:18.116792 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode367e739_45d9_4c71_82fa_ecda02da3277.slice/crio-9165416a79a8d14934c00fc8e00a91ffd697d205964c3585f55278b965651da9 WatchSource:0}: Error finding container 9165416a79a8d14934c00fc8e00a91ffd697d205964c3585f55278b965651da9: Status 404 returned error can't find the container with id 9165416a79a8d14934c00fc8e00a91ffd697d205964c3585f55278b965651da9 Jan 29 16:25:18 crc kubenswrapper[4714]: I0129 16:25:18.442964 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"e367e739-45d9-4c71-82fa-ecda02da3277","Type":"ContainerStarted","Data":"9165416a79a8d14934c00fc8e00a91ffd697d205964c3585f55278b965651da9"} Jan 29 16:25:18 crc kubenswrapper[4714]: I0129 16:25:18.444459 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" event={"ID":"2779e724-225f-4a5f-9e2c-3b05fe08dff2","Type":"ContainerStarted","Data":"cec1cafa23793d2e5f4bd3af8e35a522a2c8cc5c802408ae2ad9896bd189471e"} Jan 29 16:25:18 crc kubenswrapper[4714]: I0129 16:25:18.444638 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" Jan 29 16:25:18 crc kubenswrapper[4714]: I0129 16:25:18.445594 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"e27f02c1-a7d5-4d49-838b-df5445720a07","Type":"ContainerStarted","Data":"37a7ed7cc71d5ba4399190ff48b8e2d70a326be5e9ad5c8773900669dfc3740e"} Jan 29 16:25:18 crc kubenswrapper[4714]: I0129 16:25:18.446793 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"f8d336f2-b190-4e32-be3a-27fbf0e50a06","Type":"ContainerStarted","Data":"2730c5f20c68259dc37f00e3d986e43810a9d5fe85207a2c704b5656855e553b"} Jan 29 16:25:18 crc kubenswrapper[4714]: I0129 16:25:18.470431 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" podStartSLOduration=2.273415463 podStartE2EDuration="5.470401892s" podCreationTimestamp="2026-01-29 16:25:13 +0000 UTC" firstStartedPulling="2026-01-29 16:25:14.290221467 +0000 UTC m=+920.810722587" lastFinishedPulling="2026-01-29 16:25:17.487207906 +0000 UTC m=+924.007709016" observedRunningTime="2026-01-29 16:25:18.467824657 +0000 UTC m=+924.988325777" watchObservedRunningTime="2026-01-29 16:25:18.470401892 +0000 UTC m=+924.990903022" Jan 29 16:25:23 crc kubenswrapper[4714]: I0129 16:25:23.872593 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" Jan 29 16:25:27 crc kubenswrapper[4714]: I0129 16:25:27.366356 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/memcached-0"] Jan 29 16:25:27 crc kubenswrapper[4714]: I0129 16:25:27.367745 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/memcached-0" Jan 29 16:25:27 crc kubenswrapper[4714]: I0129 16:25:27.369672 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"memcached-config-data" Jan 29 16:25:27 crc kubenswrapper[4714]: I0129 16:25:27.370149 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"memcached-memcached-dockercfg-drt8f" Jan 29 16:25:27 crc kubenswrapper[4714]: I0129 16:25:27.380583 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Jan 29 16:25:27 crc kubenswrapper[4714]: I0129 16:25:27.516126 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea-kolla-config\") pod \"memcached-0\" (UID: \"d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea\") " pod="cinder-kuttl-tests/memcached-0" Jan 29 16:25:27 crc kubenswrapper[4714]: I0129 16:25:27.516185 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea-config-data\") pod \"memcached-0\" (UID: \"d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea\") " pod="cinder-kuttl-tests/memcached-0" Jan 29 16:25:27 crc kubenswrapper[4714]: I0129 16:25:27.516222 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkmzp\" (UniqueName: \"kubernetes.io/projected/d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea-kube-api-access-vkmzp\") pod \"memcached-0\" (UID: \"d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea\") " pod="cinder-kuttl-tests/memcached-0" Jan 29 16:25:27 crc kubenswrapper[4714]: I0129 16:25:27.617797 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea-kolla-config\") pod \"memcached-0\" (UID: \"d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea\") " pod="cinder-kuttl-tests/memcached-0" Jan 29 16:25:27 crc kubenswrapper[4714]: I0129 16:25:27.617892 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea-config-data\") pod \"memcached-0\" (UID: \"d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea\") " pod="cinder-kuttl-tests/memcached-0" Jan 29 16:25:27 crc kubenswrapper[4714]: I0129 16:25:27.618012 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkmzp\" (UniqueName: \"kubernetes.io/projected/d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea-kube-api-access-vkmzp\") pod \"memcached-0\" (UID: \"d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea\") " pod="cinder-kuttl-tests/memcached-0" Jan 29 16:25:27 crc kubenswrapper[4714]: I0129 16:25:27.618870 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea-config-data\") pod \"memcached-0\" (UID: \"d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea\") " pod="cinder-kuttl-tests/memcached-0" Jan 29 16:25:27 crc kubenswrapper[4714]: I0129 16:25:27.619245 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea-kolla-config\") pod \"memcached-0\" (UID: \"d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea\") " pod="cinder-kuttl-tests/memcached-0" Jan 29 16:25:27 crc kubenswrapper[4714]: I0129 16:25:27.657209 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkmzp\" (UniqueName: \"kubernetes.io/projected/d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea-kube-api-access-vkmzp\") pod \"memcached-0\" (UID: \"d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea\") " pod="cinder-kuttl-tests/memcached-0" Jan 29 16:25:27 crc kubenswrapper[4714]: I0129 16:25:27.684075 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/memcached-0" Jan 29 16:25:27 crc kubenswrapper[4714]: I0129 16:25:27.952818 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-794kb"] Jan 29 16:25:27 crc kubenswrapper[4714]: I0129 16:25:27.962120 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-794kb" Jan 29 16:25:27 crc kubenswrapper[4714]: I0129 16:25:27.965262 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-mqv6m" Jan 29 16:25:27 crc kubenswrapper[4714]: I0129 16:25:27.972754 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-794kb"] Jan 29 16:25:28 crc kubenswrapper[4714]: I0129 16:25:28.028851 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbp85\" (UniqueName: \"kubernetes.io/projected/f3272d6a-aac2-4e20-b996-28fc1980cd2e-kube-api-access-vbp85\") pod \"rabbitmq-cluster-operator-index-794kb\" (UID: \"f3272d6a-aac2-4e20-b996-28fc1980cd2e\") " pod="openstack-operators/rabbitmq-cluster-operator-index-794kb" Jan 29 16:25:28 crc kubenswrapper[4714]: I0129 16:25:28.131240 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbp85\" (UniqueName: \"kubernetes.io/projected/f3272d6a-aac2-4e20-b996-28fc1980cd2e-kube-api-access-vbp85\") pod \"rabbitmq-cluster-operator-index-794kb\" (UID: \"f3272d6a-aac2-4e20-b996-28fc1980cd2e\") " pod="openstack-operators/rabbitmq-cluster-operator-index-794kb" Jan 29 16:25:28 crc kubenswrapper[4714]: I0129 16:25:28.148763 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbp85\" (UniqueName: \"kubernetes.io/projected/f3272d6a-aac2-4e20-b996-28fc1980cd2e-kube-api-access-vbp85\") pod \"rabbitmq-cluster-operator-index-794kb\" (UID: \"f3272d6a-aac2-4e20-b996-28fc1980cd2e\") " pod="openstack-operators/rabbitmq-cluster-operator-index-794kb" Jan 29 16:25:28 crc kubenswrapper[4714]: I0129 16:25:28.208410 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Jan 29 16:25:28 crc kubenswrapper[4714]: I0129 16:25:28.305385 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-794kb" Jan 29 16:25:28 crc kubenswrapper[4714]: I0129 16:25:28.527960 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"e367e739-45d9-4c71-82fa-ecda02da3277","Type":"ContainerStarted","Data":"ec556421e74871bfa9c85c53ff648a1a0691767027648a2904e427bc8f75360f"} Jan 29 16:25:28 crc kubenswrapper[4714]: I0129 16:25:28.533444 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/memcached-0" event={"ID":"d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea","Type":"ContainerStarted","Data":"5fcd1b55c77976e4d94c390473639a6b02e3c4a2129659d96d1e68f62ca74a39"} Jan 29 16:25:28 crc kubenswrapper[4714]: I0129 16:25:28.805826 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-794kb"] Jan 29 16:25:29 crc kubenswrapper[4714]: I0129 16:25:29.550578 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-794kb" event={"ID":"f3272d6a-aac2-4e20-b996-28fc1980cd2e","Type":"ContainerStarted","Data":"dc124fe5adeea1043a217db158112150ae74a2240fb76a192af1e43bb85ff258"} Jan 29 16:25:29 crc kubenswrapper[4714]: I0129 16:25:29.551955 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"e27f02c1-a7d5-4d49-838b-df5445720a07","Type":"ContainerStarted","Data":"2c8a621ff05274e374040ac76aa490fcc2858782c45b93e547a0c03c210a4530"} Jan 29 16:25:30 crc kubenswrapper[4714]: I0129 16:25:30.560817 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"f8d336f2-b190-4e32-be3a-27fbf0e50a06","Type":"ContainerStarted","Data":"7b624009e8962fd057296e2a9f997c5b0aab61c294b85ef5b94d41ebe8dd89e7"} Jan 29 16:25:32 crc kubenswrapper[4714]: I0129 16:25:32.160289 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-794kb"] Jan 29 16:25:32 crc kubenswrapper[4714]: I0129 16:25:32.746155 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-ddf2f"] Jan 29 16:25:32 crc kubenswrapper[4714]: I0129 16:25:32.746808 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-ddf2f" Jan 29 16:25:32 crc kubenswrapper[4714]: I0129 16:25:32.769064 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-ddf2f"] Jan 29 16:25:32 crc kubenswrapper[4714]: I0129 16:25:32.786101 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v42s4\" (UniqueName: \"kubernetes.io/projected/e6d50b97-e5e2-426e-b881-dfb2077c0838-kube-api-access-v42s4\") pod \"rabbitmq-cluster-operator-index-ddf2f\" (UID: \"e6d50b97-e5e2-426e-b881-dfb2077c0838\") " pod="openstack-operators/rabbitmq-cluster-operator-index-ddf2f" Jan 29 16:25:32 crc kubenswrapper[4714]: I0129 16:25:32.887205 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v42s4\" (UniqueName: \"kubernetes.io/projected/e6d50b97-e5e2-426e-b881-dfb2077c0838-kube-api-access-v42s4\") pod \"rabbitmq-cluster-operator-index-ddf2f\" (UID: \"e6d50b97-e5e2-426e-b881-dfb2077c0838\") " pod="openstack-operators/rabbitmq-cluster-operator-index-ddf2f" Jan 29 16:25:32 crc kubenswrapper[4714]: I0129 16:25:32.929547 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v42s4\" (UniqueName: \"kubernetes.io/projected/e6d50b97-e5e2-426e-b881-dfb2077c0838-kube-api-access-v42s4\") pod \"rabbitmq-cluster-operator-index-ddf2f\" (UID: \"e6d50b97-e5e2-426e-b881-dfb2077c0838\") " pod="openstack-operators/rabbitmq-cluster-operator-index-ddf2f" Jan 29 16:25:33 crc kubenswrapper[4714]: I0129 16:25:33.082598 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-ddf2f" Jan 29 16:25:33 crc kubenswrapper[4714]: I0129 16:25:33.536047 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-ddf2f"] Jan 29 16:25:33 crc kubenswrapper[4714]: W0129 16:25:33.546103 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6d50b97_e5e2_426e_b881_dfb2077c0838.slice/crio-74d0d3468956b4f9354638adf80a846fa26535a7e74e370eeb721012c70bb9d7 WatchSource:0}: Error finding container 74d0d3468956b4f9354638adf80a846fa26535a7e74e370eeb721012c70bb9d7: Status 404 returned error can't find the container with id 74d0d3468956b4f9354638adf80a846fa26535a7e74e370eeb721012c70bb9d7 Jan 29 16:25:33 crc kubenswrapper[4714]: I0129 16:25:33.578686 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-ddf2f" event={"ID":"e6d50b97-e5e2-426e-b881-dfb2077c0838","Type":"ContainerStarted","Data":"74d0d3468956b4f9354638adf80a846fa26535a7e74e370eeb721012c70bb9d7"} Jan 29 16:25:33 crc kubenswrapper[4714]: I0129 16:25:33.585478 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/memcached-0" event={"ID":"d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea","Type":"ContainerStarted","Data":"5244ecc09d3c74f3a195f77af7759f9d242093032f09d5208383b4619dda295d"} Jan 29 16:25:33 crc kubenswrapper[4714]: I0129 16:25:33.585619 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/memcached-0" Jan 29 16:25:33 crc kubenswrapper[4714]: I0129 16:25:33.604518 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/memcached-0" podStartSLOduration=1.663254696 podStartE2EDuration="6.604500575s" podCreationTimestamp="2026-01-29 16:25:27 +0000 UTC" firstStartedPulling="2026-01-29 16:25:28.241387713 +0000 UTC m=+934.761888833" lastFinishedPulling="2026-01-29 16:25:33.182633592 +0000 UTC m=+939.703134712" observedRunningTime="2026-01-29 16:25:33.601494507 +0000 UTC m=+940.121995637" watchObservedRunningTime="2026-01-29 16:25:33.604500575 +0000 UTC m=+940.125001695" Jan 29 16:25:35 crc kubenswrapper[4714]: I0129 16:25:35.607096 4714 generic.go:334] "Generic (PLEG): container finished" podID="e27f02c1-a7d5-4d49-838b-df5445720a07" containerID="2c8a621ff05274e374040ac76aa490fcc2858782c45b93e547a0c03c210a4530" exitCode=0 Jan 29 16:25:35 crc kubenswrapper[4714]: I0129 16:25:35.607174 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"e27f02c1-a7d5-4d49-838b-df5445720a07","Type":"ContainerDied","Data":"2c8a621ff05274e374040ac76aa490fcc2858782c45b93e547a0c03c210a4530"} Jan 29 16:25:35 crc kubenswrapper[4714]: I0129 16:25:35.610149 4714 generic.go:334] "Generic (PLEG): container finished" podID="f8d336f2-b190-4e32-be3a-27fbf0e50a06" containerID="7b624009e8962fd057296e2a9f997c5b0aab61c294b85ef5b94d41ebe8dd89e7" exitCode=0 Jan 29 16:25:35 crc kubenswrapper[4714]: I0129 16:25:35.610243 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"f8d336f2-b190-4e32-be3a-27fbf0e50a06","Type":"ContainerDied","Data":"7b624009e8962fd057296e2a9f997c5b0aab61c294b85ef5b94d41ebe8dd89e7"} Jan 29 16:25:35 crc kubenswrapper[4714]: I0129 16:25:35.612375 4714 generic.go:334] "Generic (PLEG): container finished" podID="e367e739-45d9-4c71-82fa-ecda02da3277" containerID="ec556421e74871bfa9c85c53ff648a1a0691767027648a2904e427bc8f75360f" exitCode=0 Jan 29 16:25:35 crc kubenswrapper[4714]: I0129 16:25:35.612414 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"e367e739-45d9-4c71-82fa-ecda02da3277","Type":"ContainerDied","Data":"ec556421e74871bfa9c85c53ff648a1a0691767027648a2904e427bc8f75360f"} Jan 29 16:25:36 crc kubenswrapper[4714]: I0129 16:25:36.620627 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"e27f02c1-a7d5-4d49-838b-df5445720a07","Type":"ContainerStarted","Data":"b027a08741bc7af4590803c43740e388681a5f71be01b0ea2974650f0e3e74fc"} Jan 29 16:25:36 crc kubenswrapper[4714]: I0129 16:25:36.622452 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-ddf2f" event={"ID":"e6d50b97-e5e2-426e-b881-dfb2077c0838","Type":"ContainerStarted","Data":"f6b5016e56cc05cedf6153ecd27ad5475f41f09c64ae9c2dcbdb602e86bd0445"} Jan 29 16:25:36 crc kubenswrapper[4714]: I0129 16:25:36.624576 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"f8d336f2-b190-4e32-be3a-27fbf0e50a06","Type":"ContainerStarted","Data":"83c3eb1ecb12cd8202e4a2f2b14330aa7199092be01cfe200e08827657c44a8b"} Jan 29 16:25:36 crc kubenswrapper[4714]: I0129 16:25:36.626748 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-794kb" event={"ID":"f3272d6a-aac2-4e20-b996-28fc1980cd2e","Type":"ContainerStarted","Data":"d4336bc35a9f3b5419f76f6a79f33203bd10152a030a0548f63621c695133a4c"} Jan 29 16:25:36 crc kubenswrapper[4714]: I0129 16:25:36.626849 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-794kb" podUID="f3272d6a-aac2-4e20-b996-28fc1980cd2e" containerName="registry-server" containerID="cri-o://d4336bc35a9f3b5419f76f6a79f33203bd10152a030a0548f63621c695133a4c" gracePeriod=2 Jan 29 16:25:36 crc kubenswrapper[4714]: I0129 16:25:36.629267 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"e367e739-45d9-4c71-82fa-ecda02da3277","Type":"ContainerStarted","Data":"6653bc0073fbdde40468fb06b953c1302846763f2888418b4fc21509ef3915d2"} Jan 29 16:25:36 crc kubenswrapper[4714]: I0129 16:25:36.646524 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/openstack-galera-0" podStartSLOduration=10.499287299 podStartE2EDuration="20.646506011s" podCreationTimestamp="2026-01-29 16:25:16 +0000 UTC" firstStartedPulling="2026-01-29 16:25:18.062804216 +0000 UTC m=+924.583305336" lastFinishedPulling="2026-01-29 16:25:28.210022928 +0000 UTC m=+934.730524048" observedRunningTime="2026-01-29 16:25:36.641401232 +0000 UTC m=+943.161902352" watchObservedRunningTime="2026-01-29 16:25:36.646506011 +0000 UTC m=+943.167007131" Jan 29 16:25:36 crc kubenswrapper[4714]: I0129 16:25:36.662470 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-ddf2f" podStartSLOduration=2.3995356 podStartE2EDuration="4.662447836s" podCreationTimestamp="2026-01-29 16:25:32 +0000 UTC" firstStartedPulling="2026-01-29 16:25:33.547423399 +0000 UTC m=+940.067924519" lastFinishedPulling="2026-01-29 16:25:35.810335635 +0000 UTC m=+942.330836755" observedRunningTime="2026-01-29 16:25:36.657842721 +0000 UTC m=+943.178343841" watchObservedRunningTime="2026-01-29 16:25:36.662447836 +0000 UTC m=+943.182948956" Jan 29 16:25:36 crc kubenswrapper[4714]: I0129 16:25:36.682124 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/openstack-galera-1" podStartSLOduration=9.28406374 podStartE2EDuration="20.68211016s" podCreationTimestamp="2026-01-29 16:25:16 +0000 UTC" firstStartedPulling="2026-01-29 16:25:18.110086206 +0000 UTC m=+924.630587316" lastFinishedPulling="2026-01-29 16:25:29.508132616 +0000 UTC m=+936.028633736" observedRunningTime="2026-01-29 16:25:36.678418582 +0000 UTC m=+943.198919712" watchObservedRunningTime="2026-01-29 16:25:36.68211016 +0000 UTC m=+943.202611280" Jan 29 16:25:36 crc kubenswrapper[4714]: I0129 16:25:36.701298 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/openstack-galera-2" podStartSLOduration=10.672422561 podStartE2EDuration="20.701283239s" podCreationTimestamp="2026-01-29 16:25:16 +0000 UTC" firstStartedPulling="2026-01-29 16:25:18.119177351 +0000 UTC m=+924.639678471" lastFinishedPulling="2026-01-29 16:25:28.148038029 +0000 UTC m=+934.668539149" observedRunningTime="2026-01-29 16:25:36.697686524 +0000 UTC m=+943.218187644" watchObservedRunningTime="2026-01-29 16:25:36.701283239 +0000 UTC m=+943.221784359" Jan 29 16:25:36 crc kubenswrapper[4714]: I0129 16:25:36.722466 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-794kb" podStartSLOduration=2.7445488559999998 podStartE2EDuration="9.722446597s" podCreationTimestamp="2026-01-29 16:25:27 +0000 UTC" firstStartedPulling="2026-01-29 16:25:28.817355214 +0000 UTC m=+935.337856334" lastFinishedPulling="2026-01-29 16:25:35.795252955 +0000 UTC m=+942.315754075" observedRunningTime="2026-01-29 16:25:36.718541633 +0000 UTC m=+943.239042753" watchObservedRunningTime="2026-01-29 16:25:36.722446597 +0000 UTC m=+943.242947717" Jan 29 16:25:37 crc kubenswrapper[4714]: I0129 16:25:37.077522 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-794kb" Jan 29 16:25:37 crc kubenswrapper[4714]: I0129 16:25:37.154291 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbp85\" (UniqueName: \"kubernetes.io/projected/f3272d6a-aac2-4e20-b996-28fc1980cd2e-kube-api-access-vbp85\") pod \"f3272d6a-aac2-4e20-b996-28fc1980cd2e\" (UID: \"f3272d6a-aac2-4e20-b996-28fc1980cd2e\") " Jan 29 16:25:37 crc kubenswrapper[4714]: I0129 16:25:37.160047 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3272d6a-aac2-4e20-b996-28fc1980cd2e-kube-api-access-vbp85" (OuterVolumeSpecName: "kube-api-access-vbp85") pod "f3272d6a-aac2-4e20-b996-28fc1980cd2e" (UID: "f3272d6a-aac2-4e20-b996-28fc1980cd2e"). InnerVolumeSpecName "kube-api-access-vbp85". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:25:37 crc kubenswrapper[4714]: I0129 16:25:37.255953 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbp85\" (UniqueName: \"kubernetes.io/projected/f3272d6a-aac2-4e20-b996-28fc1980cd2e-kube-api-access-vbp85\") on node \"crc\" DevicePath \"\"" Jan 29 16:25:37 crc kubenswrapper[4714]: I0129 16:25:37.616508 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:37 crc kubenswrapper[4714]: I0129 16:25:37.616817 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:25:37 crc kubenswrapper[4714]: I0129 16:25:37.635782 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:37 crc kubenswrapper[4714]: I0129 16:25:37.635829 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:25:37 crc kubenswrapper[4714]: I0129 16:25:37.637309 4714 generic.go:334] "Generic (PLEG): container finished" podID="f3272d6a-aac2-4e20-b996-28fc1980cd2e" containerID="d4336bc35a9f3b5419f76f6a79f33203bd10152a030a0548f63621c695133a4c" exitCode=0 Jan 29 16:25:37 crc kubenswrapper[4714]: I0129 16:25:37.637352 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-794kb" Jan 29 16:25:37 crc kubenswrapper[4714]: I0129 16:25:37.637352 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-794kb" event={"ID":"f3272d6a-aac2-4e20-b996-28fc1980cd2e","Type":"ContainerDied","Data":"d4336bc35a9f3b5419f76f6a79f33203bd10152a030a0548f63621c695133a4c"} Jan 29 16:25:37 crc kubenswrapper[4714]: I0129 16:25:37.637384 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-794kb" event={"ID":"f3272d6a-aac2-4e20-b996-28fc1980cd2e","Type":"ContainerDied","Data":"dc124fe5adeea1043a217db158112150ae74a2240fb76a192af1e43bb85ff258"} Jan 29 16:25:37 crc kubenswrapper[4714]: I0129 16:25:37.637400 4714 scope.go:117] "RemoveContainer" containerID="d4336bc35a9f3b5419f76f6a79f33203bd10152a030a0548f63621c695133a4c" Jan 29 16:25:37 crc kubenswrapper[4714]: I0129 16:25:37.650731 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:37 crc kubenswrapper[4714]: I0129 16:25:37.650770 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:37 crc kubenswrapper[4714]: I0129 16:25:37.655980 4714 scope.go:117] "RemoveContainer" containerID="d4336bc35a9f3b5419f76f6a79f33203bd10152a030a0548f63621c695133a4c" Jan 29 16:25:37 crc kubenswrapper[4714]: E0129 16:25:37.656384 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4336bc35a9f3b5419f76f6a79f33203bd10152a030a0548f63621c695133a4c\": container with ID starting with d4336bc35a9f3b5419f76f6a79f33203bd10152a030a0548f63621c695133a4c not found: ID does not exist" containerID="d4336bc35a9f3b5419f76f6a79f33203bd10152a030a0548f63621c695133a4c" Jan 29 16:25:37 crc kubenswrapper[4714]: I0129 16:25:37.656422 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4336bc35a9f3b5419f76f6a79f33203bd10152a030a0548f63621c695133a4c"} err="failed to get container status \"d4336bc35a9f3b5419f76f6a79f33203bd10152a030a0548f63621c695133a4c\": rpc error: code = NotFound desc = could not find container \"d4336bc35a9f3b5419f76f6a79f33203bd10152a030a0548f63621c695133a4c\": container with ID starting with d4336bc35a9f3b5419f76f6a79f33203bd10152a030a0548f63621c695133a4c not found: ID does not exist" Jan 29 16:25:37 crc kubenswrapper[4714]: I0129 16:25:37.668274 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-794kb"] Jan 29 16:25:37 crc kubenswrapper[4714]: I0129 16:25:37.671981 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-794kb"] Jan 29 16:25:38 crc kubenswrapper[4714]: I0129 16:25:38.191128 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3272d6a-aac2-4e20-b996-28fc1980cd2e" path="/var/lib/kubelet/pods/f3272d6a-aac2-4e20-b996-28fc1980cd2e/volumes" Jan 29 16:25:39 crc kubenswrapper[4714]: I0129 16:25:39.153088 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lmqpk"] Jan 29 16:25:39 crc kubenswrapper[4714]: E0129 16:25:39.153296 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3272d6a-aac2-4e20-b996-28fc1980cd2e" containerName="registry-server" Jan 29 16:25:39 crc kubenswrapper[4714]: I0129 16:25:39.153307 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3272d6a-aac2-4e20-b996-28fc1980cd2e" containerName="registry-server" Jan 29 16:25:39 crc kubenswrapper[4714]: I0129 16:25:39.153417 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3272d6a-aac2-4e20-b996-28fc1980cd2e" containerName="registry-server" Jan 29 16:25:39 crc kubenswrapper[4714]: I0129 16:25:39.154160 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lmqpk" Jan 29 16:25:39 crc kubenswrapper[4714]: I0129 16:25:39.169094 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lmqpk"] Jan 29 16:25:39 crc kubenswrapper[4714]: I0129 16:25:39.302338 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fppdq\" (UniqueName: \"kubernetes.io/projected/230dbe44-24bd-4a95-9f71-7ee36bb74cce-kube-api-access-fppdq\") pod \"community-operators-lmqpk\" (UID: \"230dbe44-24bd-4a95-9f71-7ee36bb74cce\") " pod="openshift-marketplace/community-operators-lmqpk" Jan 29 16:25:39 crc kubenswrapper[4714]: I0129 16:25:39.302420 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/230dbe44-24bd-4a95-9f71-7ee36bb74cce-catalog-content\") pod \"community-operators-lmqpk\" (UID: \"230dbe44-24bd-4a95-9f71-7ee36bb74cce\") " pod="openshift-marketplace/community-operators-lmqpk" Jan 29 16:25:39 crc kubenswrapper[4714]: I0129 16:25:39.302490 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/230dbe44-24bd-4a95-9f71-7ee36bb74cce-utilities\") pod \"community-operators-lmqpk\" (UID: \"230dbe44-24bd-4a95-9f71-7ee36bb74cce\") " pod="openshift-marketplace/community-operators-lmqpk" Jan 29 16:25:39 crc kubenswrapper[4714]: I0129 16:25:39.404261 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/230dbe44-24bd-4a95-9f71-7ee36bb74cce-catalog-content\") pod \"community-operators-lmqpk\" (UID: \"230dbe44-24bd-4a95-9f71-7ee36bb74cce\") " pod="openshift-marketplace/community-operators-lmqpk" Jan 29 16:25:39 crc kubenswrapper[4714]: I0129 16:25:39.404711 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/230dbe44-24bd-4a95-9f71-7ee36bb74cce-utilities\") pod \"community-operators-lmqpk\" (UID: \"230dbe44-24bd-4a95-9f71-7ee36bb74cce\") " pod="openshift-marketplace/community-operators-lmqpk" Jan 29 16:25:39 crc kubenswrapper[4714]: I0129 16:25:39.404751 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fppdq\" (UniqueName: \"kubernetes.io/projected/230dbe44-24bd-4a95-9f71-7ee36bb74cce-kube-api-access-fppdq\") pod \"community-operators-lmqpk\" (UID: \"230dbe44-24bd-4a95-9f71-7ee36bb74cce\") " pod="openshift-marketplace/community-operators-lmqpk" Jan 29 16:25:39 crc kubenswrapper[4714]: I0129 16:25:39.404833 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/230dbe44-24bd-4a95-9f71-7ee36bb74cce-catalog-content\") pod \"community-operators-lmqpk\" (UID: \"230dbe44-24bd-4a95-9f71-7ee36bb74cce\") " pod="openshift-marketplace/community-operators-lmqpk" Jan 29 16:25:39 crc kubenswrapper[4714]: I0129 16:25:39.405354 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/230dbe44-24bd-4a95-9f71-7ee36bb74cce-utilities\") pod \"community-operators-lmqpk\" (UID: \"230dbe44-24bd-4a95-9f71-7ee36bb74cce\") " pod="openshift-marketplace/community-operators-lmqpk" Jan 29 16:25:39 crc kubenswrapper[4714]: I0129 16:25:39.426127 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fppdq\" (UniqueName: \"kubernetes.io/projected/230dbe44-24bd-4a95-9f71-7ee36bb74cce-kube-api-access-fppdq\") pod \"community-operators-lmqpk\" (UID: \"230dbe44-24bd-4a95-9f71-7ee36bb74cce\") " pod="openshift-marketplace/community-operators-lmqpk" Jan 29 16:25:39 crc kubenswrapper[4714]: I0129 16:25:39.470738 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lmqpk" Jan 29 16:25:39 crc kubenswrapper[4714]: I0129 16:25:39.919053 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lmqpk"] Jan 29 16:25:39 crc kubenswrapper[4714]: W0129 16:25:39.925213 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod230dbe44_24bd_4a95_9f71_7ee36bb74cce.slice/crio-b963773a81de472dcc9c53a373aee4fe097bee647f3f7cf04a8f2ef907468156 WatchSource:0}: Error finding container b963773a81de472dcc9c53a373aee4fe097bee647f3f7cf04a8f2ef907468156: Status 404 returned error can't find the container with id b963773a81de472dcc9c53a373aee4fe097bee647f3f7cf04a8f2ef907468156 Jan 29 16:25:40 crc kubenswrapper[4714]: I0129 16:25:40.676547 4714 generic.go:334] "Generic (PLEG): container finished" podID="230dbe44-24bd-4a95-9f71-7ee36bb74cce" containerID="5b730a9dfe53ac182298170cbbd5116f09f6dc7bc63b9980a70b7d90c7997752" exitCode=0 Jan 29 16:25:40 crc kubenswrapper[4714]: I0129 16:25:40.676604 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lmqpk" event={"ID":"230dbe44-24bd-4a95-9f71-7ee36bb74cce","Type":"ContainerDied","Data":"5b730a9dfe53ac182298170cbbd5116f09f6dc7bc63b9980a70b7d90c7997752"} Jan 29 16:25:40 crc kubenswrapper[4714]: I0129 16:25:40.676843 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lmqpk" event={"ID":"230dbe44-24bd-4a95-9f71-7ee36bb74cce","Type":"ContainerStarted","Data":"b963773a81de472dcc9c53a373aee4fe097bee647f3f7cf04a8f2ef907468156"} Jan 29 16:25:42 crc kubenswrapper[4714]: I0129 16:25:42.685139 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/memcached-0" Jan 29 16:25:43 crc kubenswrapper[4714]: I0129 16:25:43.086059 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-ddf2f" Jan 29 16:25:43 crc kubenswrapper[4714]: I0129 16:25:43.086119 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-ddf2f" Jan 29 16:25:43 crc kubenswrapper[4714]: I0129 16:25:43.127860 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-ddf2f" Jan 29 16:25:43 crc kubenswrapper[4714]: I0129 16:25:43.713550 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-ddf2f" Jan 29 16:25:46 crc kubenswrapper[4714]: I0129 16:25:46.711226 4714 generic.go:334] "Generic (PLEG): container finished" podID="230dbe44-24bd-4a95-9f71-7ee36bb74cce" containerID="d22cb237e69d1d81bb2cf37fadc3bb5651ed34bb7ad52432494b2ea7410d53d9" exitCode=0 Jan 29 16:25:46 crc kubenswrapper[4714]: I0129 16:25:46.711324 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lmqpk" event={"ID":"230dbe44-24bd-4a95-9f71-7ee36bb74cce","Type":"ContainerDied","Data":"d22cb237e69d1d81bb2cf37fadc3bb5651ed34bb7ad52432494b2ea7410d53d9"} Jan 29 16:25:47 crc kubenswrapper[4714]: I0129 16:25:47.206197 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s"] Jan 29 16:25:47 crc kubenswrapper[4714]: I0129 16:25:47.208404 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s" Jan 29 16:25:47 crc kubenswrapper[4714]: I0129 16:25:47.213336 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-hwqbr" Jan 29 16:25:47 crc kubenswrapper[4714]: I0129 16:25:47.219880 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s"] Jan 29 16:25:47 crc kubenswrapper[4714]: I0129 16:25:47.305715 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/74070831-862a-4d0a-83b0-4e3d64891601-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s\" (UID: \"74070831-862a-4d0a-83b0-4e3d64891601\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s" Jan 29 16:25:47 crc kubenswrapper[4714]: I0129 16:25:47.305801 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/74070831-862a-4d0a-83b0-4e3d64891601-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s\" (UID: \"74070831-862a-4d0a-83b0-4e3d64891601\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s" Jan 29 16:25:47 crc kubenswrapper[4714]: I0129 16:25:47.305860 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwwng\" (UniqueName: \"kubernetes.io/projected/74070831-862a-4d0a-83b0-4e3d64891601-kube-api-access-dwwng\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s\" (UID: \"74070831-862a-4d0a-83b0-4e3d64891601\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s" Jan 29 16:25:47 crc kubenswrapper[4714]: I0129 16:25:47.407413 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwwng\" (UniqueName: \"kubernetes.io/projected/74070831-862a-4d0a-83b0-4e3d64891601-kube-api-access-dwwng\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s\" (UID: \"74070831-862a-4d0a-83b0-4e3d64891601\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s" Jan 29 16:25:47 crc kubenswrapper[4714]: I0129 16:25:47.407732 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/74070831-862a-4d0a-83b0-4e3d64891601-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s\" (UID: \"74070831-862a-4d0a-83b0-4e3d64891601\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s" Jan 29 16:25:47 crc kubenswrapper[4714]: I0129 16:25:47.407766 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/74070831-862a-4d0a-83b0-4e3d64891601-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s\" (UID: \"74070831-862a-4d0a-83b0-4e3d64891601\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s" Jan 29 16:25:47 crc kubenswrapper[4714]: I0129 16:25:47.408217 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/74070831-862a-4d0a-83b0-4e3d64891601-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s\" (UID: \"74070831-862a-4d0a-83b0-4e3d64891601\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s" Jan 29 16:25:47 crc kubenswrapper[4714]: I0129 16:25:47.408251 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/74070831-862a-4d0a-83b0-4e3d64891601-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s\" (UID: \"74070831-862a-4d0a-83b0-4e3d64891601\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s" Jan 29 16:25:47 crc kubenswrapper[4714]: I0129 16:25:47.426301 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwwng\" (UniqueName: \"kubernetes.io/projected/74070831-862a-4d0a-83b0-4e3d64891601-kube-api-access-dwwng\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s\" (UID: \"74070831-862a-4d0a-83b0-4e3d64891601\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s" Jan 29 16:25:47 crc kubenswrapper[4714]: I0129 16:25:47.526751 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s" Jan 29 16:25:47 crc kubenswrapper[4714]: I0129 16:25:47.727806 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lmqpk" event={"ID":"230dbe44-24bd-4a95-9f71-7ee36bb74cce","Type":"ContainerStarted","Data":"486c03c8de207535cc686dfcaa8ec86e8491b149b1078748b5ef7e236bd5cbd3"} Jan 29 16:25:47 crc kubenswrapper[4714]: I0129 16:25:47.743720 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lmqpk" podStartSLOduration=2.201220656 podStartE2EDuration="8.743697709s" podCreationTimestamp="2026-01-29 16:25:39 +0000 UTC" firstStartedPulling="2026-01-29 16:25:40.678004475 +0000 UTC m=+947.198505595" lastFinishedPulling="2026-01-29 16:25:47.220481528 +0000 UTC m=+953.740982648" observedRunningTime="2026-01-29 16:25:47.741448953 +0000 UTC m=+954.261950073" watchObservedRunningTime="2026-01-29 16:25:47.743697709 +0000 UTC m=+954.264198829" Jan 29 16:25:47 crc kubenswrapper[4714]: I0129 16:25:47.773924 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:47 crc kubenswrapper[4714]: I0129 16:25:47.849485 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:25:47 crc kubenswrapper[4714]: I0129 16:25:47.954950 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s"] Jan 29 16:25:48 crc kubenswrapper[4714]: I0129 16:25:48.734694 4714 generic.go:334] "Generic (PLEG): container finished" podID="74070831-862a-4d0a-83b0-4e3d64891601" containerID="08be91d5ade94d67396b725df7d3290e5e0b4eed8f678b830b5a24bf0aefb822" exitCode=0 Jan 29 16:25:48 crc kubenswrapper[4714]: I0129 16:25:48.734794 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s" event={"ID":"74070831-862a-4d0a-83b0-4e3d64891601","Type":"ContainerDied","Data":"08be91d5ade94d67396b725df7d3290e5e0b4eed8f678b830b5a24bf0aefb822"} Jan 29 16:25:48 crc kubenswrapper[4714]: I0129 16:25:48.734843 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s" event={"ID":"74070831-862a-4d0a-83b0-4e3d64891601","Type":"ContainerStarted","Data":"52b0391a8d42deb2c4b8c1aff420808b658d4de8de74c3e4fd16389cd7b50f5b"} Jan 29 16:25:49 crc kubenswrapper[4714]: I0129 16:25:49.471471 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lmqpk" Jan 29 16:25:49 crc kubenswrapper[4714]: I0129 16:25:49.471580 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lmqpk" Jan 29 16:25:49 crc kubenswrapper[4714]: I0129 16:25:49.524567 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lmqpk" Jan 29 16:25:50 crc kubenswrapper[4714]: I0129 16:25:50.747649 4714 generic.go:334] "Generic (PLEG): container finished" podID="74070831-862a-4d0a-83b0-4e3d64891601" containerID="9a1454161993efe4d3d18d4c054f92025122ffa043acc6e49d820a2c93adec47" exitCode=0 Jan 29 16:25:50 crc kubenswrapper[4714]: I0129 16:25:50.747810 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s" event={"ID":"74070831-862a-4d0a-83b0-4e3d64891601","Type":"ContainerDied","Data":"9a1454161993efe4d3d18d4c054f92025122ffa043acc6e49d820a2c93adec47"} Jan 29 16:25:51 crc kubenswrapper[4714]: I0129 16:25:51.757860 4714 generic.go:334] "Generic (PLEG): container finished" podID="74070831-862a-4d0a-83b0-4e3d64891601" containerID="bf336482d3003324aec4b339b20443d78b6e477a15e7d92b43bd42b82e826811" exitCode=0 Jan 29 16:25:51 crc kubenswrapper[4714]: I0129 16:25:51.757987 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s" event={"ID":"74070831-862a-4d0a-83b0-4e3d64891601","Type":"ContainerDied","Data":"bf336482d3003324aec4b339b20443d78b6e477a15e7d92b43bd42b82e826811"} Jan 29 16:25:53 crc kubenswrapper[4714]: I0129 16:25:53.113005 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s" Jan 29 16:25:53 crc kubenswrapper[4714]: I0129 16:25:53.190658 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/74070831-862a-4d0a-83b0-4e3d64891601-bundle\") pod \"74070831-862a-4d0a-83b0-4e3d64891601\" (UID: \"74070831-862a-4d0a-83b0-4e3d64891601\") " Jan 29 16:25:53 crc kubenswrapper[4714]: I0129 16:25:53.190698 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/74070831-862a-4d0a-83b0-4e3d64891601-util\") pod \"74070831-862a-4d0a-83b0-4e3d64891601\" (UID: \"74070831-862a-4d0a-83b0-4e3d64891601\") " Jan 29 16:25:53 crc kubenswrapper[4714]: I0129 16:25:53.190737 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwwng\" (UniqueName: \"kubernetes.io/projected/74070831-862a-4d0a-83b0-4e3d64891601-kube-api-access-dwwng\") pod \"74070831-862a-4d0a-83b0-4e3d64891601\" (UID: \"74070831-862a-4d0a-83b0-4e3d64891601\") " Jan 29 16:25:53 crc kubenswrapper[4714]: I0129 16:25:53.191549 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74070831-862a-4d0a-83b0-4e3d64891601-bundle" (OuterVolumeSpecName: "bundle") pod "74070831-862a-4d0a-83b0-4e3d64891601" (UID: "74070831-862a-4d0a-83b0-4e3d64891601"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:25:53 crc kubenswrapper[4714]: I0129 16:25:53.198037 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74070831-862a-4d0a-83b0-4e3d64891601-kube-api-access-dwwng" (OuterVolumeSpecName: "kube-api-access-dwwng") pod "74070831-862a-4d0a-83b0-4e3d64891601" (UID: "74070831-862a-4d0a-83b0-4e3d64891601"). InnerVolumeSpecName "kube-api-access-dwwng". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:25:53 crc kubenswrapper[4714]: I0129 16:25:53.209996 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74070831-862a-4d0a-83b0-4e3d64891601-util" (OuterVolumeSpecName: "util") pod "74070831-862a-4d0a-83b0-4e3d64891601" (UID: "74070831-862a-4d0a-83b0-4e3d64891601"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:25:53 crc kubenswrapper[4714]: I0129 16:25:53.292578 4714 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/74070831-862a-4d0a-83b0-4e3d64891601-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 16:25:53 crc kubenswrapper[4714]: I0129 16:25:53.292623 4714 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/74070831-862a-4d0a-83b0-4e3d64891601-util\") on node \"crc\" DevicePath \"\"" Jan 29 16:25:53 crc kubenswrapper[4714]: I0129 16:25:53.292633 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwwng\" (UniqueName: \"kubernetes.io/projected/74070831-862a-4d0a-83b0-4e3d64891601-kube-api-access-dwwng\") on node \"crc\" DevicePath \"\"" Jan 29 16:25:53 crc kubenswrapper[4714]: I0129 16:25:53.778281 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s" event={"ID":"74070831-862a-4d0a-83b0-4e3d64891601","Type":"ContainerDied","Data":"52b0391a8d42deb2c4b8c1aff420808b658d4de8de74c3e4fd16389cd7b50f5b"} Jan 29 16:25:53 crc kubenswrapper[4714]: I0129 16:25:53.778323 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52b0391a8d42deb2c4b8c1aff420808b658d4de8de74c3e4fd16389cd7b50f5b" Jan 29 16:25:53 crc kubenswrapper[4714]: I0129 16:25:53.778364 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s" Jan 29 16:25:54 crc kubenswrapper[4714]: I0129 16:25:54.955706 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wdwq5"] Jan 29 16:25:54 crc kubenswrapper[4714]: E0129 16:25:54.956128 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74070831-862a-4d0a-83b0-4e3d64891601" containerName="util" Jan 29 16:25:54 crc kubenswrapper[4714]: I0129 16:25:54.956156 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="74070831-862a-4d0a-83b0-4e3d64891601" containerName="util" Jan 29 16:25:54 crc kubenswrapper[4714]: E0129 16:25:54.956178 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74070831-862a-4d0a-83b0-4e3d64891601" containerName="pull" Jan 29 16:25:54 crc kubenswrapper[4714]: I0129 16:25:54.956188 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="74070831-862a-4d0a-83b0-4e3d64891601" containerName="pull" Jan 29 16:25:54 crc kubenswrapper[4714]: E0129 16:25:54.956216 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74070831-862a-4d0a-83b0-4e3d64891601" containerName="extract" Jan 29 16:25:54 crc kubenswrapper[4714]: I0129 16:25:54.956227 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="74070831-862a-4d0a-83b0-4e3d64891601" containerName="extract" Jan 29 16:25:54 crc kubenswrapper[4714]: I0129 16:25:54.956421 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="74070831-862a-4d0a-83b0-4e3d64891601" containerName="extract" Jan 29 16:25:54 crc kubenswrapper[4714]: I0129 16:25:54.957535 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wdwq5" Jan 29 16:25:54 crc kubenswrapper[4714]: I0129 16:25:54.968872 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdwq5"] Jan 29 16:25:55 crc kubenswrapper[4714]: I0129 16:25:55.116641 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmm88\" (UniqueName: \"kubernetes.io/projected/8c12ad14-f878-42a1-a168-bad4026ec2dd-kube-api-access-tmm88\") pod \"redhat-marketplace-wdwq5\" (UID: \"8c12ad14-f878-42a1-a168-bad4026ec2dd\") " pod="openshift-marketplace/redhat-marketplace-wdwq5" Jan 29 16:25:55 crc kubenswrapper[4714]: I0129 16:25:55.116786 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c12ad14-f878-42a1-a168-bad4026ec2dd-utilities\") pod \"redhat-marketplace-wdwq5\" (UID: \"8c12ad14-f878-42a1-a168-bad4026ec2dd\") " pod="openshift-marketplace/redhat-marketplace-wdwq5" Jan 29 16:25:55 crc kubenswrapper[4714]: I0129 16:25:55.116856 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c12ad14-f878-42a1-a168-bad4026ec2dd-catalog-content\") pod \"redhat-marketplace-wdwq5\" (UID: \"8c12ad14-f878-42a1-a168-bad4026ec2dd\") " pod="openshift-marketplace/redhat-marketplace-wdwq5" Jan 29 16:25:55 crc kubenswrapper[4714]: I0129 16:25:55.218061 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c12ad14-f878-42a1-a168-bad4026ec2dd-catalog-content\") pod \"redhat-marketplace-wdwq5\" (UID: \"8c12ad14-f878-42a1-a168-bad4026ec2dd\") " pod="openshift-marketplace/redhat-marketplace-wdwq5" Jan 29 16:25:55 crc kubenswrapper[4714]: I0129 16:25:55.218118 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmm88\" (UniqueName: \"kubernetes.io/projected/8c12ad14-f878-42a1-a168-bad4026ec2dd-kube-api-access-tmm88\") pod \"redhat-marketplace-wdwq5\" (UID: \"8c12ad14-f878-42a1-a168-bad4026ec2dd\") " pod="openshift-marketplace/redhat-marketplace-wdwq5" Jan 29 16:25:55 crc kubenswrapper[4714]: I0129 16:25:55.218182 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c12ad14-f878-42a1-a168-bad4026ec2dd-utilities\") pod \"redhat-marketplace-wdwq5\" (UID: \"8c12ad14-f878-42a1-a168-bad4026ec2dd\") " pod="openshift-marketplace/redhat-marketplace-wdwq5" Jan 29 16:25:55 crc kubenswrapper[4714]: I0129 16:25:55.218668 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c12ad14-f878-42a1-a168-bad4026ec2dd-utilities\") pod \"redhat-marketplace-wdwq5\" (UID: \"8c12ad14-f878-42a1-a168-bad4026ec2dd\") " pod="openshift-marketplace/redhat-marketplace-wdwq5" Jan 29 16:25:55 crc kubenswrapper[4714]: I0129 16:25:55.218878 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c12ad14-f878-42a1-a168-bad4026ec2dd-catalog-content\") pod \"redhat-marketplace-wdwq5\" (UID: \"8c12ad14-f878-42a1-a168-bad4026ec2dd\") " pod="openshift-marketplace/redhat-marketplace-wdwq5" Jan 29 16:25:55 crc kubenswrapper[4714]: I0129 16:25:55.238025 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmm88\" (UniqueName: \"kubernetes.io/projected/8c12ad14-f878-42a1-a168-bad4026ec2dd-kube-api-access-tmm88\") pod \"redhat-marketplace-wdwq5\" (UID: \"8c12ad14-f878-42a1-a168-bad4026ec2dd\") " pod="openshift-marketplace/redhat-marketplace-wdwq5" Jan 29 16:25:55 crc kubenswrapper[4714]: I0129 16:25:55.273607 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wdwq5" Jan 29 16:25:55 crc kubenswrapper[4714]: I0129 16:25:55.723256 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdwq5"] Jan 29 16:25:55 crc kubenswrapper[4714]: W0129 16:25:55.730415 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c12ad14_f878_42a1_a168_bad4026ec2dd.slice/crio-e9d23b1dec5222eaf00f0f2fac6279153030320fe12205a5c55c774a975165f4 WatchSource:0}: Error finding container e9d23b1dec5222eaf00f0f2fac6279153030320fe12205a5c55c774a975165f4: Status 404 returned error can't find the container with id e9d23b1dec5222eaf00f0f2fac6279153030320fe12205a5c55c774a975165f4 Jan 29 16:25:55 crc kubenswrapper[4714]: I0129 16:25:55.795129 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdwq5" event={"ID":"8c12ad14-f878-42a1-a168-bad4026ec2dd","Type":"ContainerStarted","Data":"e9d23b1dec5222eaf00f0f2fac6279153030320fe12205a5c55c774a975165f4"} Jan 29 16:25:56 crc kubenswrapper[4714]: I0129 16:25:56.365051 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/root-account-create-update-2fh2r"] Jan 29 16:25:56 crc kubenswrapper[4714]: I0129 16:25:56.365754 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/root-account-create-update-2fh2r" Jan 29 16:25:56 crc kubenswrapper[4714]: I0129 16:25:56.368521 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"openstack-mariadb-root-db-secret" Jan 29 16:25:56 crc kubenswrapper[4714]: I0129 16:25:56.381968 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/root-account-create-update-2fh2r"] Jan 29 16:25:56 crc kubenswrapper[4714]: I0129 16:25:56.443546 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn2vl\" (UniqueName: \"kubernetes.io/projected/c94e6f97-6224-46d2-b406-f5d02a596cb7-kube-api-access-jn2vl\") pod \"root-account-create-update-2fh2r\" (UID: \"c94e6f97-6224-46d2-b406-f5d02a596cb7\") " pod="cinder-kuttl-tests/root-account-create-update-2fh2r" Jan 29 16:25:56 crc kubenswrapper[4714]: I0129 16:25:56.444196 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c94e6f97-6224-46d2-b406-f5d02a596cb7-operator-scripts\") pod \"root-account-create-update-2fh2r\" (UID: \"c94e6f97-6224-46d2-b406-f5d02a596cb7\") " pod="cinder-kuttl-tests/root-account-create-update-2fh2r" Jan 29 16:25:56 crc kubenswrapper[4714]: I0129 16:25:56.545601 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn2vl\" (UniqueName: \"kubernetes.io/projected/c94e6f97-6224-46d2-b406-f5d02a596cb7-kube-api-access-jn2vl\") pod \"root-account-create-update-2fh2r\" (UID: \"c94e6f97-6224-46d2-b406-f5d02a596cb7\") " pod="cinder-kuttl-tests/root-account-create-update-2fh2r" Jan 29 16:25:56 crc kubenswrapper[4714]: I0129 16:25:56.545702 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c94e6f97-6224-46d2-b406-f5d02a596cb7-operator-scripts\") pod \"root-account-create-update-2fh2r\" (UID: \"c94e6f97-6224-46d2-b406-f5d02a596cb7\") " pod="cinder-kuttl-tests/root-account-create-update-2fh2r" Jan 29 16:25:56 crc kubenswrapper[4714]: I0129 16:25:56.546397 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c94e6f97-6224-46d2-b406-f5d02a596cb7-operator-scripts\") pod \"root-account-create-update-2fh2r\" (UID: \"c94e6f97-6224-46d2-b406-f5d02a596cb7\") " pod="cinder-kuttl-tests/root-account-create-update-2fh2r" Jan 29 16:25:56 crc kubenswrapper[4714]: I0129 16:25:56.571905 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn2vl\" (UniqueName: \"kubernetes.io/projected/c94e6f97-6224-46d2-b406-f5d02a596cb7-kube-api-access-jn2vl\") pod \"root-account-create-update-2fh2r\" (UID: \"c94e6f97-6224-46d2-b406-f5d02a596cb7\") " pod="cinder-kuttl-tests/root-account-create-update-2fh2r" Jan 29 16:25:56 crc kubenswrapper[4714]: I0129 16:25:56.681264 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/root-account-create-update-2fh2r" Jan 29 16:25:56 crc kubenswrapper[4714]: I0129 16:25:56.805644 4714 generic.go:334] "Generic (PLEG): container finished" podID="8c12ad14-f878-42a1-a168-bad4026ec2dd" containerID="efa3dccd677eb8ead685e31efa78ca9b7337f378a73a59452840851ca329f89c" exitCode=0 Jan 29 16:25:56 crc kubenswrapper[4714]: I0129 16:25:56.806017 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdwq5" event={"ID":"8c12ad14-f878-42a1-a168-bad4026ec2dd","Type":"ContainerDied","Data":"efa3dccd677eb8ead685e31efa78ca9b7337f378a73a59452840851ca329f89c"} Jan 29 16:25:57 crc kubenswrapper[4714]: I0129 16:25:57.133437 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/root-account-create-update-2fh2r"] Jan 29 16:25:57 crc kubenswrapper[4714]: E0129 16:25:57.458750 4714 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 16:25:57 crc kubenswrapper[4714]: E0129 16:25:57.459272 4714 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tmm88,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-wdwq5_openshift-marketplace(8c12ad14-f878-42a1-a168-bad4026ec2dd): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 16:25:57 crc kubenswrapper[4714]: E0129 16:25:57.461320 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:25:57 crc kubenswrapper[4714]: I0129 16:25:57.728448 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/openstack-galera-2" podUID="e367e739-45d9-4c71-82fa-ecda02da3277" containerName="galera" probeResult="failure" output=< Jan 29 16:25:57 crc kubenswrapper[4714]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Jan 29 16:25:57 crc kubenswrapper[4714]: > Jan 29 16:25:57 crc kubenswrapper[4714]: I0129 16:25:57.822381 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/root-account-create-update-2fh2r" event={"ID":"c94e6f97-6224-46d2-b406-f5d02a596cb7","Type":"ContainerStarted","Data":"ae1e8fd69fe054dba679bc3d816ff2486311c7fd767f49bb3c77b8a2f9da9054"} Jan 29 16:25:57 crc kubenswrapper[4714]: I0129 16:25:57.822662 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/root-account-create-update-2fh2r" event={"ID":"c94e6f97-6224-46d2-b406-f5d02a596cb7","Type":"ContainerStarted","Data":"21ee325ce54c33324918e25586a05eb462c8b87e131df00ae9fd89812d284160"} Jan 29 16:25:57 crc kubenswrapper[4714]: E0129 16:25:57.823953 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:25:57 crc kubenswrapper[4714]: I0129 16:25:57.889763 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/root-account-create-update-2fh2r" podStartSLOduration=1.889744627 podStartE2EDuration="1.889744627s" podCreationTimestamp="2026-01-29 16:25:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:25:57.886582265 +0000 UTC m=+964.407083385" watchObservedRunningTime="2026-01-29 16:25:57.889744627 +0000 UTC m=+964.410245747" Jan 29 16:25:59 crc kubenswrapper[4714]: I0129 16:25:59.512270 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lmqpk" Jan 29 16:25:59 crc kubenswrapper[4714]: I0129 16:25:59.836662 4714 generic.go:334] "Generic (PLEG): container finished" podID="c94e6f97-6224-46d2-b406-f5d02a596cb7" containerID="ae1e8fd69fe054dba679bc3d816ff2486311c7fd767f49bb3c77b8a2f9da9054" exitCode=0 Jan 29 16:25:59 crc kubenswrapper[4714]: I0129 16:25:59.836715 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/root-account-create-update-2fh2r" event={"ID":"c94e6f97-6224-46d2-b406-f5d02a596cb7","Type":"ContainerDied","Data":"ae1e8fd69fe054dba679bc3d816ff2486311c7fd767f49bb3c77b8a2f9da9054"} Jan 29 16:26:00 crc kubenswrapper[4714]: I0129 16:26:00.123489 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:26:00 crc kubenswrapper[4714]: I0129 16:26:00.219161 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:26:01 crc kubenswrapper[4714]: I0129 16:26:01.175046 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/root-account-create-update-2fh2r" Jan 29 16:26:01 crc kubenswrapper[4714]: I0129 16:26:01.334327 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jn2vl\" (UniqueName: \"kubernetes.io/projected/c94e6f97-6224-46d2-b406-f5d02a596cb7-kube-api-access-jn2vl\") pod \"c94e6f97-6224-46d2-b406-f5d02a596cb7\" (UID: \"c94e6f97-6224-46d2-b406-f5d02a596cb7\") " Jan 29 16:26:01 crc kubenswrapper[4714]: I0129 16:26:01.334516 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c94e6f97-6224-46d2-b406-f5d02a596cb7-operator-scripts\") pod \"c94e6f97-6224-46d2-b406-f5d02a596cb7\" (UID: \"c94e6f97-6224-46d2-b406-f5d02a596cb7\") " Jan 29 16:26:01 crc kubenswrapper[4714]: I0129 16:26:01.335041 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c94e6f97-6224-46d2-b406-f5d02a596cb7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c94e6f97-6224-46d2-b406-f5d02a596cb7" (UID: "c94e6f97-6224-46d2-b406-f5d02a596cb7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:26:01 crc kubenswrapper[4714]: I0129 16:26:01.345606 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c94e6f97-6224-46d2-b406-f5d02a596cb7-kube-api-access-jn2vl" (OuterVolumeSpecName: "kube-api-access-jn2vl") pod "c94e6f97-6224-46d2-b406-f5d02a596cb7" (UID: "c94e6f97-6224-46d2-b406-f5d02a596cb7"). InnerVolumeSpecName "kube-api-access-jn2vl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:26:01 crc kubenswrapper[4714]: I0129 16:26:01.436364 4714 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c94e6f97-6224-46d2-b406-f5d02a596cb7-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:26:01 crc kubenswrapper[4714]: I0129 16:26:01.436448 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jn2vl\" (UniqueName: \"kubernetes.io/projected/c94e6f97-6224-46d2-b406-f5d02a596cb7-kube-api-access-jn2vl\") on node \"crc\" DevicePath \"\"" Jan 29 16:26:01 crc kubenswrapper[4714]: I0129 16:26:01.850005 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/root-account-create-update-2fh2r" event={"ID":"c94e6f97-6224-46d2-b406-f5d02a596cb7","Type":"ContainerDied","Data":"21ee325ce54c33324918e25586a05eb462c8b87e131df00ae9fd89812d284160"} Jan 29 16:26:01 crc kubenswrapper[4714]: I0129 16:26:01.850052 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21ee325ce54c33324918e25586a05eb462c8b87e131df00ae9fd89812d284160" Jan 29 16:26:01 crc kubenswrapper[4714]: I0129 16:26:01.850055 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/root-account-create-update-2fh2r" Jan 29 16:26:02 crc kubenswrapper[4714]: I0129 16:26:02.221656 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:26:02 crc kubenswrapper[4714]: I0129 16:26:02.339134 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:26:04 crc kubenswrapper[4714]: I0129 16:26:04.342708 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lmqpk"] Jan 29 16:26:04 crc kubenswrapper[4714]: I0129 16:26:04.343069 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lmqpk" podUID="230dbe44-24bd-4a95-9f71-7ee36bb74cce" containerName="registry-server" containerID="cri-o://486c03c8de207535cc686dfcaa8ec86e8491b149b1078748b5ef7e236bd5cbd3" gracePeriod=2 Jan 29 16:26:04 crc kubenswrapper[4714]: I0129 16:26:04.873161 4714 generic.go:334] "Generic (PLEG): container finished" podID="230dbe44-24bd-4a95-9f71-7ee36bb74cce" containerID="486c03c8de207535cc686dfcaa8ec86e8491b149b1078748b5ef7e236bd5cbd3" exitCode=0 Jan 29 16:26:04 crc kubenswrapper[4714]: I0129 16:26:04.873223 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lmqpk" event={"ID":"230dbe44-24bd-4a95-9f71-7ee36bb74cce","Type":"ContainerDied","Data":"486c03c8de207535cc686dfcaa8ec86e8491b149b1078748b5ef7e236bd5cbd3"} Jan 29 16:26:05 crc kubenswrapper[4714]: I0129 16:26:05.325242 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lmqpk" Jan 29 16:26:05 crc kubenswrapper[4714]: I0129 16:26:05.392183 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/230dbe44-24bd-4a95-9f71-7ee36bb74cce-utilities\") pod \"230dbe44-24bd-4a95-9f71-7ee36bb74cce\" (UID: \"230dbe44-24bd-4a95-9f71-7ee36bb74cce\") " Jan 29 16:26:05 crc kubenswrapper[4714]: I0129 16:26:05.392839 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fppdq\" (UniqueName: \"kubernetes.io/projected/230dbe44-24bd-4a95-9f71-7ee36bb74cce-kube-api-access-fppdq\") pod \"230dbe44-24bd-4a95-9f71-7ee36bb74cce\" (UID: \"230dbe44-24bd-4a95-9f71-7ee36bb74cce\") " Jan 29 16:26:05 crc kubenswrapper[4714]: I0129 16:26:05.393070 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/230dbe44-24bd-4a95-9f71-7ee36bb74cce-catalog-content\") pod \"230dbe44-24bd-4a95-9f71-7ee36bb74cce\" (UID: \"230dbe44-24bd-4a95-9f71-7ee36bb74cce\") " Jan 29 16:26:05 crc kubenswrapper[4714]: I0129 16:26:05.393188 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/230dbe44-24bd-4a95-9f71-7ee36bb74cce-utilities" (OuterVolumeSpecName: "utilities") pod "230dbe44-24bd-4a95-9f71-7ee36bb74cce" (UID: "230dbe44-24bd-4a95-9f71-7ee36bb74cce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:26:05 crc kubenswrapper[4714]: I0129 16:26:05.400454 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/230dbe44-24bd-4a95-9f71-7ee36bb74cce-kube-api-access-fppdq" (OuterVolumeSpecName: "kube-api-access-fppdq") pod "230dbe44-24bd-4a95-9f71-7ee36bb74cce" (UID: "230dbe44-24bd-4a95-9f71-7ee36bb74cce"). InnerVolumeSpecName "kube-api-access-fppdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:26:05 crc kubenswrapper[4714]: I0129 16:26:05.445747 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/230dbe44-24bd-4a95-9f71-7ee36bb74cce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "230dbe44-24bd-4a95-9f71-7ee36bb74cce" (UID: "230dbe44-24bd-4a95-9f71-7ee36bb74cce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:26:05 crc kubenswrapper[4714]: I0129 16:26:05.495308 4714 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/230dbe44-24bd-4a95-9f71-7ee36bb74cce-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:26:05 crc kubenswrapper[4714]: I0129 16:26:05.495344 4714 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/230dbe44-24bd-4a95-9f71-7ee36bb74cce-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:26:05 crc kubenswrapper[4714]: I0129 16:26:05.495354 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fppdq\" (UniqueName: \"kubernetes.io/projected/230dbe44-24bd-4a95-9f71-7ee36bb74cce-kube-api-access-fppdq\") on node \"crc\" DevicePath \"\"" Jan 29 16:26:05 crc kubenswrapper[4714]: I0129 16:26:05.881039 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lmqpk" event={"ID":"230dbe44-24bd-4a95-9f71-7ee36bb74cce","Type":"ContainerDied","Data":"b963773a81de472dcc9c53a373aee4fe097bee647f3f7cf04a8f2ef907468156"} Jan 29 16:26:05 crc kubenswrapper[4714]: I0129 16:26:05.881095 4714 scope.go:117] "RemoveContainer" containerID="486c03c8de207535cc686dfcaa8ec86e8491b149b1078748b5ef7e236bd5cbd3" Jan 29 16:26:05 crc kubenswrapper[4714]: I0129 16:26:05.881105 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lmqpk" Jan 29 16:26:05 crc kubenswrapper[4714]: I0129 16:26:05.916524 4714 scope.go:117] "RemoveContainer" containerID="d22cb237e69d1d81bb2cf37fadc3bb5651ed34bb7ad52432494b2ea7410d53d9" Jan 29 16:26:05 crc kubenswrapper[4714]: I0129 16:26:05.917588 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lmqpk"] Jan 29 16:26:05 crc kubenswrapper[4714]: I0129 16:26:05.924701 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lmqpk"] Jan 29 16:26:05 crc kubenswrapper[4714]: I0129 16:26:05.937353 4714 scope.go:117] "RemoveContainer" containerID="5b730a9dfe53ac182298170cbbd5116f09f6dc7bc63b9980a70b7d90c7997752" Jan 29 16:26:06 crc kubenswrapper[4714]: I0129 16:26:06.195168 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="230dbe44-24bd-4a95-9f71-7ee36bb74cce" path="/var/lib/kubelet/pods/230dbe44-24bd-4a95-9f71-7ee36bb74cce/volumes" Jan 29 16:26:08 crc kubenswrapper[4714]: I0129 16:26:08.153477 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-xprqx"] Jan 29 16:26:08 crc kubenswrapper[4714]: E0129 16:26:08.153792 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="230dbe44-24bd-4a95-9f71-7ee36bb74cce" containerName="registry-server" Jan 29 16:26:08 crc kubenswrapper[4714]: I0129 16:26:08.153808 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="230dbe44-24bd-4a95-9f71-7ee36bb74cce" containerName="registry-server" Jan 29 16:26:08 crc kubenswrapper[4714]: E0129 16:26:08.153825 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="230dbe44-24bd-4a95-9f71-7ee36bb74cce" containerName="extract-utilities" Jan 29 16:26:08 crc kubenswrapper[4714]: I0129 16:26:08.153834 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="230dbe44-24bd-4a95-9f71-7ee36bb74cce" containerName="extract-utilities" Jan 29 16:26:08 crc kubenswrapper[4714]: E0129 16:26:08.153852 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c94e6f97-6224-46d2-b406-f5d02a596cb7" containerName="mariadb-account-create-update" Jan 29 16:26:08 crc kubenswrapper[4714]: I0129 16:26:08.153860 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="c94e6f97-6224-46d2-b406-f5d02a596cb7" containerName="mariadb-account-create-update" Jan 29 16:26:08 crc kubenswrapper[4714]: E0129 16:26:08.153873 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="230dbe44-24bd-4a95-9f71-7ee36bb74cce" containerName="extract-content" Jan 29 16:26:08 crc kubenswrapper[4714]: I0129 16:26:08.153881 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="230dbe44-24bd-4a95-9f71-7ee36bb74cce" containerName="extract-content" Jan 29 16:26:08 crc kubenswrapper[4714]: I0129 16:26:08.154043 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="c94e6f97-6224-46d2-b406-f5d02a596cb7" containerName="mariadb-account-create-update" Jan 29 16:26:08 crc kubenswrapper[4714]: I0129 16:26:08.154062 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="230dbe44-24bd-4a95-9f71-7ee36bb74cce" containerName="registry-server" Jan 29 16:26:08 crc kubenswrapper[4714]: I0129 16:26:08.154597 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-xprqx" Jan 29 16:26:08 crc kubenswrapper[4714]: I0129 16:26:08.159976 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-d5fnd" Jan 29 16:26:08 crc kubenswrapper[4714]: I0129 16:26:08.171434 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-xprqx"] Jan 29 16:26:08 crc kubenswrapper[4714]: I0129 16:26:08.240463 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2fkn\" (UniqueName: \"kubernetes.io/projected/12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88-kube-api-access-p2fkn\") pod \"rabbitmq-cluster-operator-779fc9694b-xprqx\" (UID: \"12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-xprqx" Jan 29 16:26:08 crc kubenswrapper[4714]: I0129 16:26:08.342121 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2fkn\" (UniqueName: \"kubernetes.io/projected/12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88-kube-api-access-p2fkn\") pod \"rabbitmq-cluster-operator-779fc9694b-xprqx\" (UID: \"12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-xprqx" Jan 29 16:26:08 crc kubenswrapper[4714]: I0129 16:26:08.361017 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2fkn\" (UniqueName: \"kubernetes.io/projected/12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88-kube-api-access-p2fkn\") pod \"rabbitmq-cluster-operator-779fc9694b-xprqx\" (UID: \"12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-xprqx" Jan 29 16:26:08 crc kubenswrapper[4714]: I0129 16:26:08.486771 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-xprqx" Jan 29 16:26:08 crc kubenswrapper[4714]: I0129 16:26:08.938628 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-xprqx"] Jan 29 16:26:09 crc kubenswrapper[4714]: I0129 16:26:09.919830 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-xprqx" event={"ID":"12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88","Type":"ContainerStarted","Data":"9cef1a74877ad1a74c47eaad2e3c11a8681670f01de34b07c07c5549def07a12"} Jan 29 16:26:11 crc kubenswrapper[4714]: E0129 16:26:11.568307 4714 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 16:26:11 crc kubenswrapper[4714]: E0129 16:26:11.569403 4714 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tmm88,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-wdwq5_openshift-marketplace(8c12ad14-f878-42a1-a168-bad4026ec2dd): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 16:26:11 crc kubenswrapper[4714]: E0129 16:26:11.570662 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:26:12 crc kubenswrapper[4714]: I0129 16:26:12.943163 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-xprqx" event={"ID":"12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88","Type":"ContainerStarted","Data":"0c6b038c874feac00a0ea4a05fbd431f7014a4ed8fc71291c23ec972425e646d"} Jan 29 16:26:12 crc kubenswrapper[4714]: I0129 16:26:12.965972 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-xprqx" podStartSLOduration=1.9700317840000001 podStartE2EDuration="4.965894368s" podCreationTimestamp="2026-01-29 16:26:08 +0000 UTC" firstStartedPulling="2026-01-29 16:26:08.948000927 +0000 UTC m=+975.468502047" lastFinishedPulling="2026-01-29 16:26:11.943863511 +0000 UTC m=+978.464364631" observedRunningTime="2026-01-29 16:26:12.964282593 +0000 UTC m=+979.484783783" watchObservedRunningTime="2026-01-29 16:26:12.965894368 +0000 UTC m=+979.486395528" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.740287 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.741681 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.743430 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"rabbitmq-server-dockercfg-4pzhl" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.743546 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"rabbitmq-plugins-conf" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.744189 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"rabbitmq-erlang-cookie" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.744266 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"rabbitmq-server-conf" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.744445 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"rabbitmq-default-user" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.751840 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.854997 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pgjf\" (UniqueName: \"kubernetes.io/projected/55e23ac1-a89b-4689-a17d-bee875f7783e-kube-api-access-8pgjf\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.855045 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/55e23ac1-a89b-4689-a17d-bee875f7783e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.855128 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/55e23ac1-a89b-4689-a17d-bee875f7783e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.855209 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/55e23ac1-a89b-4689-a17d-bee875f7783e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.855337 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/55e23ac1-a89b-4689-a17d-bee875f7783e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.855358 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/55e23ac1-a89b-4689-a17d-bee875f7783e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.855379 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/55e23ac1-a89b-4689-a17d-bee875f7783e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.855416 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-21c03d9a-bbed-44e7-9b38-49f8915c54d2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-21c03d9a-bbed-44e7-9b38-49f8915c54d2\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.956490 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/55e23ac1-a89b-4689-a17d-bee875f7783e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.956568 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/55e23ac1-a89b-4689-a17d-bee875f7783e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.956653 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/55e23ac1-a89b-4689-a17d-bee875f7783e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.956687 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/55e23ac1-a89b-4689-a17d-bee875f7783e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.956718 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/55e23ac1-a89b-4689-a17d-bee875f7783e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.956757 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-21c03d9a-bbed-44e7-9b38-49f8915c54d2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-21c03d9a-bbed-44e7-9b38-49f8915c54d2\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.956810 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pgjf\" (UniqueName: \"kubernetes.io/projected/55e23ac1-a89b-4689-a17d-bee875f7783e-kube-api-access-8pgjf\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.956845 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/55e23ac1-a89b-4689-a17d-bee875f7783e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.958049 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/55e23ac1-a89b-4689-a17d-bee875f7783e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.959554 4714 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.959591 4714 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-21c03d9a-bbed-44e7-9b38-49f8915c54d2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-21c03d9a-bbed-44e7-9b38-49f8915c54d2\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b382ee18eee6fccf46656a68ff47f48e9aa5ccd13f3cf6bd6d751a5365eb3cfb/globalmount\"" pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.959886 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/55e23ac1-a89b-4689-a17d-bee875f7783e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.960309 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/55e23ac1-a89b-4689-a17d-bee875f7783e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.965678 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/55e23ac1-a89b-4689-a17d-bee875f7783e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.965677 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/55e23ac1-a89b-4689-a17d-bee875f7783e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.971780 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/55e23ac1-a89b-4689-a17d-bee875f7783e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.976203 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pgjf\" (UniqueName: \"kubernetes.io/projected/55e23ac1-a89b-4689-a17d-bee875f7783e-kube-api-access-8pgjf\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:15 crc kubenswrapper[4714]: I0129 16:26:15.996383 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-21c03d9a-bbed-44e7-9b38-49f8915c54d2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-21c03d9a-bbed-44e7-9b38-49f8915c54d2\") pod \"rabbitmq-server-0\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:16 crc kubenswrapper[4714]: I0129 16:26:16.091085 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:16 crc kubenswrapper[4714]: I0129 16:26:16.514965 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Jan 29 16:26:16 crc kubenswrapper[4714]: W0129 16:26:16.520229 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55e23ac1_a89b_4689_a17d_bee875f7783e.slice/crio-299eb5904909cd50412aad25a871f9888d290b3d4f1acfe53103c96e6f05a1bc WatchSource:0}: Error finding container 299eb5904909cd50412aad25a871f9888d290b3d4f1acfe53103c96e6f05a1bc: Status 404 returned error can't find the container with id 299eb5904909cd50412aad25a871f9888d290b3d4f1acfe53103c96e6f05a1bc Jan 29 16:26:16 crc kubenswrapper[4714]: I0129 16:26:16.967118 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"55e23ac1-a89b-4689-a17d-bee875f7783e","Type":"ContainerStarted","Data":"299eb5904909cd50412aad25a871f9888d290b3d4f1acfe53103c96e6f05a1bc"} Jan 29 16:26:18 crc kubenswrapper[4714]: I0129 16:26:18.352305 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-vtc5h"] Jan 29 16:26:18 crc kubenswrapper[4714]: I0129 16:26:18.354856 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-vtc5h" Jan 29 16:26:18 crc kubenswrapper[4714]: I0129 16:26:18.358106 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-klkzg" Jan 29 16:26:18 crc kubenswrapper[4714]: I0129 16:26:18.361804 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-vtc5h"] Jan 29 16:26:18 crc kubenswrapper[4714]: I0129 16:26:18.509896 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsc27\" (UniqueName: \"kubernetes.io/projected/668764e7-6295-4275-bcc9-24b680ec685f-kube-api-access-wsc27\") pod \"keystone-operator-index-vtc5h\" (UID: \"668764e7-6295-4275-bcc9-24b680ec685f\") " pod="openstack-operators/keystone-operator-index-vtc5h" Jan 29 16:26:18 crc kubenswrapper[4714]: I0129 16:26:18.610661 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsc27\" (UniqueName: \"kubernetes.io/projected/668764e7-6295-4275-bcc9-24b680ec685f-kube-api-access-wsc27\") pod \"keystone-operator-index-vtc5h\" (UID: \"668764e7-6295-4275-bcc9-24b680ec685f\") " pod="openstack-operators/keystone-operator-index-vtc5h" Jan 29 16:26:18 crc kubenswrapper[4714]: I0129 16:26:18.637442 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsc27\" (UniqueName: \"kubernetes.io/projected/668764e7-6295-4275-bcc9-24b680ec685f-kube-api-access-wsc27\") pod \"keystone-operator-index-vtc5h\" (UID: \"668764e7-6295-4275-bcc9-24b680ec685f\") " pod="openstack-operators/keystone-operator-index-vtc5h" Jan 29 16:26:18 crc kubenswrapper[4714]: I0129 16:26:18.723582 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-vtc5h" Jan 29 16:26:18 crc kubenswrapper[4714]: I0129 16:26:18.763063 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-w58tg"] Jan 29 16:26:18 crc kubenswrapper[4714]: I0129 16:26:18.765084 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w58tg" Jan 29 16:26:18 crc kubenswrapper[4714]: I0129 16:26:18.770775 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w58tg"] Jan 29 16:26:18 crc kubenswrapper[4714]: I0129 16:26:18.914060 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db1a5ea4-712c-4ec8-aeae-c44998fa51e1-catalog-content\") pod \"certified-operators-w58tg\" (UID: \"db1a5ea4-712c-4ec8-aeae-c44998fa51e1\") " pod="openshift-marketplace/certified-operators-w58tg" Jan 29 16:26:18 crc kubenswrapper[4714]: I0129 16:26:18.914152 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db1a5ea4-712c-4ec8-aeae-c44998fa51e1-utilities\") pod \"certified-operators-w58tg\" (UID: \"db1a5ea4-712c-4ec8-aeae-c44998fa51e1\") " pod="openshift-marketplace/certified-operators-w58tg" Jan 29 16:26:18 crc kubenswrapper[4714]: I0129 16:26:18.914188 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn67x\" (UniqueName: \"kubernetes.io/projected/db1a5ea4-712c-4ec8-aeae-c44998fa51e1-kube-api-access-rn67x\") pod \"certified-operators-w58tg\" (UID: \"db1a5ea4-712c-4ec8-aeae-c44998fa51e1\") " pod="openshift-marketplace/certified-operators-w58tg" Jan 29 16:26:19 crc kubenswrapper[4714]: I0129 16:26:19.015244 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db1a5ea4-712c-4ec8-aeae-c44998fa51e1-utilities\") pod \"certified-operators-w58tg\" (UID: \"db1a5ea4-712c-4ec8-aeae-c44998fa51e1\") " pod="openshift-marketplace/certified-operators-w58tg" Jan 29 16:26:19 crc kubenswrapper[4714]: I0129 16:26:19.015290 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rn67x\" (UniqueName: \"kubernetes.io/projected/db1a5ea4-712c-4ec8-aeae-c44998fa51e1-kube-api-access-rn67x\") pod \"certified-operators-w58tg\" (UID: \"db1a5ea4-712c-4ec8-aeae-c44998fa51e1\") " pod="openshift-marketplace/certified-operators-w58tg" Jan 29 16:26:19 crc kubenswrapper[4714]: I0129 16:26:19.015374 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db1a5ea4-712c-4ec8-aeae-c44998fa51e1-catalog-content\") pod \"certified-operators-w58tg\" (UID: \"db1a5ea4-712c-4ec8-aeae-c44998fa51e1\") " pod="openshift-marketplace/certified-operators-w58tg" Jan 29 16:26:19 crc kubenswrapper[4714]: I0129 16:26:19.015909 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db1a5ea4-712c-4ec8-aeae-c44998fa51e1-utilities\") pod \"certified-operators-w58tg\" (UID: \"db1a5ea4-712c-4ec8-aeae-c44998fa51e1\") " pod="openshift-marketplace/certified-operators-w58tg" Jan 29 16:26:19 crc kubenswrapper[4714]: I0129 16:26:19.015959 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db1a5ea4-712c-4ec8-aeae-c44998fa51e1-catalog-content\") pod \"certified-operators-w58tg\" (UID: \"db1a5ea4-712c-4ec8-aeae-c44998fa51e1\") " pod="openshift-marketplace/certified-operators-w58tg" Jan 29 16:26:19 crc kubenswrapper[4714]: I0129 16:26:19.032772 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rn67x\" (UniqueName: \"kubernetes.io/projected/db1a5ea4-712c-4ec8-aeae-c44998fa51e1-kube-api-access-rn67x\") pod \"certified-operators-w58tg\" (UID: \"db1a5ea4-712c-4ec8-aeae-c44998fa51e1\") " pod="openshift-marketplace/certified-operators-w58tg" Jan 29 16:26:19 crc kubenswrapper[4714]: I0129 16:26:19.133010 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w58tg" Jan 29 16:26:19 crc kubenswrapper[4714]: I0129 16:26:19.182262 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-vtc5h"] Jan 29 16:26:19 crc kubenswrapper[4714]: W0129 16:26:19.192006 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod668764e7_6295_4275_bcc9_24b680ec685f.slice/crio-047c3adf9e0c960c54099b8f9a0a168467b6e29a2c3acff719ceb9bbe1f69c79 WatchSource:0}: Error finding container 047c3adf9e0c960c54099b8f9a0a168467b6e29a2c3acff719ceb9bbe1f69c79: Status 404 returned error can't find the container with id 047c3adf9e0c960c54099b8f9a0a168467b6e29a2c3acff719ceb9bbe1f69c79 Jan 29 16:26:19 crc kubenswrapper[4714]: I0129 16:26:19.566806 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w58tg"] Jan 29 16:26:19 crc kubenswrapper[4714]: I0129 16:26:19.987170 4714 generic.go:334] "Generic (PLEG): container finished" podID="db1a5ea4-712c-4ec8-aeae-c44998fa51e1" containerID="977047a32f415a72ab936afe2d1774146105933da01176d2aabeb71559b4afad" exitCode=0 Jan 29 16:26:19 crc kubenswrapper[4714]: I0129 16:26:19.987263 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w58tg" event={"ID":"db1a5ea4-712c-4ec8-aeae-c44998fa51e1","Type":"ContainerDied","Data":"977047a32f415a72ab936afe2d1774146105933da01176d2aabeb71559b4afad"} Jan 29 16:26:19 crc kubenswrapper[4714]: I0129 16:26:19.987486 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w58tg" event={"ID":"db1a5ea4-712c-4ec8-aeae-c44998fa51e1","Type":"ContainerStarted","Data":"fa3e0fa56a351fb40b73e185e0120625630ac7fd4a18cf99fdfb008b5ae7fadf"} Jan 29 16:26:19 crc kubenswrapper[4714]: I0129 16:26:19.989049 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-vtc5h" event={"ID":"668764e7-6295-4275-bcc9-24b680ec685f","Type":"ContainerStarted","Data":"047c3adf9e0c960c54099b8f9a0a168467b6e29a2c3acff719ceb9bbe1f69c79"} Jan 29 16:26:20 crc kubenswrapper[4714]: E0129 16:26:20.131227 4714 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 16:26:20 crc kubenswrapper[4714]: E0129 16:26:20.131479 4714 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rn67x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-w58tg_openshift-marketplace(db1a5ea4-712c-4ec8-aeae-c44998fa51e1): ErrImagePull: initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 16:26:20 crc kubenswrapper[4714]: E0129 16:26:20.132666 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/certified-operators-w58tg" podUID="db1a5ea4-712c-4ec8-aeae-c44998fa51e1" Jan 29 16:26:20 crc kubenswrapper[4714]: I0129 16:26:20.998071 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-vtc5h" event={"ID":"668764e7-6295-4275-bcc9-24b680ec685f","Type":"ContainerStarted","Data":"dc243f10cd3a85e410a8968c999f4a42fc022a8b377403e54fe6779138d67aa8"} Jan 29 16:26:21 crc kubenswrapper[4714]: I0129 16:26:21.029626 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-vtc5h" podStartSLOduration=2.06601316 podStartE2EDuration="3.029607142s" podCreationTimestamp="2026-01-29 16:26:18 +0000 UTC" firstStartedPulling="2026-01-29 16:26:19.197171482 +0000 UTC m=+985.717672612" lastFinishedPulling="2026-01-29 16:26:20.160765474 +0000 UTC m=+986.681266594" observedRunningTime="2026-01-29 16:26:21.025909219 +0000 UTC m=+987.546410339" watchObservedRunningTime="2026-01-29 16:26:21.029607142 +0000 UTC m=+987.550108262" Jan 29 16:26:21 crc kubenswrapper[4714]: E0129 16:26:21.740692 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-w58tg" podUID="db1a5ea4-712c-4ec8-aeae-c44998fa51e1" Jan 29 16:26:23 crc kubenswrapper[4714]: E0129 16:26:23.207454 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:26:27 crc kubenswrapper[4714]: I0129 16:26:27.045018 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"55e23ac1-a89b-4689-a17d-bee875f7783e","Type":"ContainerStarted","Data":"eb95fbc4965d0aeffea6fce3a32742ee3a1bc5446fd33b5341ed6d4be042d493"} Jan 29 16:26:27 crc kubenswrapper[4714]: I0129 16:26:27.844764 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:26:27 crc kubenswrapper[4714]: I0129 16:26:27.845123 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:26:28 crc kubenswrapper[4714]: I0129 16:26:28.725043 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-vtc5h" Jan 29 16:26:28 crc kubenswrapper[4714]: I0129 16:26:28.725083 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-vtc5h" Jan 29 16:26:28 crc kubenswrapper[4714]: I0129 16:26:28.756802 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-vtc5h" Jan 29 16:26:29 crc kubenswrapper[4714]: I0129 16:26:29.082747 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-vtc5h" Jan 29 16:26:30 crc kubenswrapper[4714]: I0129 16:26:30.791826 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg"] Jan 29 16:26:30 crc kubenswrapper[4714]: I0129 16:26:30.793175 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg" Jan 29 16:26:30 crc kubenswrapper[4714]: I0129 16:26:30.840521 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-hwqbr" Jan 29 16:26:30 crc kubenswrapper[4714]: I0129 16:26:30.855288 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg"] Jan 29 16:26:30 crc kubenswrapper[4714]: I0129 16:26:30.888591 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dd598a7a-34ba-4392-908b-c18d89648bb5-util\") pod \"b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg\" (UID: \"dd598a7a-34ba-4392-908b-c18d89648bb5\") " pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg" Jan 29 16:26:30 crc kubenswrapper[4714]: I0129 16:26:30.888642 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlxp5\" (UniqueName: \"kubernetes.io/projected/dd598a7a-34ba-4392-908b-c18d89648bb5-kube-api-access-wlxp5\") pod \"b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg\" (UID: \"dd598a7a-34ba-4392-908b-c18d89648bb5\") " pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg" Jan 29 16:26:30 crc kubenswrapper[4714]: I0129 16:26:30.888875 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dd598a7a-34ba-4392-908b-c18d89648bb5-bundle\") pod \"b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg\" (UID: \"dd598a7a-34ba-4392-908b-c18d89648bb5\") " pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg" Jan 29 16:26:30 crc kubenswrapper[4714]: I0129 16:26:30.990275 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dd598a7a-34ba-4392-908b-c18d89648bb5-bundle\") pod \"b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg\" (UID: \"dd598a7a-34ba-4392-908b-c18d89648bb5\") " pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg" Jan 29 16:26:30 crc kubenswrapper[4714]: I0129 16:26:30.991233 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dd598a7a-34ba-4392-908b-c18d89648bb5-bundle\") pod \"b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg\" (UID: \"dd598a7a-34ba-4392-908b-c18d89648bb5\") " pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg" Jan 29 16:26:30 crc kubenswrapper[4714]: I0129 16:26:30.991343 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dd598a7a-34ba-4392-908b-c18d89648bb5-util\") pod \"b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg\" (UID: \"dd598a7a-34ba-4392-908b-c18d89648bb5\") " pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg" Jan 29 16:26:30 crc kubenswrapper[4714]: I0129 16:26:30.990575 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dd598a7a-34ba-4392-908b-c18d89648bb5-util\") pod \"b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg\" (UID: \"dd598a7a-34ba-4392-908b-c18d89648bb5\") " pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg" Jan 29 16:26:30 crc kubenswrapper[4714]: I0129 16:26:30.991494 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlxp5\" (UniqueName: \"kubernetes.io/projected/dd598a7a-34ba-4392-908b-c18d89648bb5-kube-api-access-wlxp5\") pod \"b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg\" (UID: \"dd598a7a-34ba-4392-908b-c18d89648bb5\") " pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg" Jan 29 16:26:31 crc kubenswrapper[4714]: I0129 16:26:31.011309 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlxp5\" (UniqueName: \"kubernetes.io/projected/dd598a7a-34ba-4392-908b-c18d89648bb5-kube-api-access-wlxp5\") pod \"b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg\" (UID: \"dd598a7a-34ba-4392-908b-c18d89648bb5\") " pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg" Jan 29 16:26:31 crc kubenswrapper[4714]: I0129 16:26:31.163857 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg" Jan 29 16:26:31 crc kubenswrapper[4714]: I0129 16:26:31.543395 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg"] Jan 29 16:26:32 crc kubenswrapper[4714]: I0129 16:26:32.074609 4714 generic.go:334] "Generic (PLEG): container finished" podID="dd598a7a-34ba-4392-908b-c18d89648bb5" containerID="cdd9e5087dbe100b4200dc045cc6536b7bd5644b604c1ad48cd724f12116a2d5" exitCode=0 Jan 29 16:26:32 crc kubenswrapper[4714]: I0129 16:26:32.074722 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg" event={"ID":"dd598a7a-34ba-4392-908b-c18d89648bb5","Type":"ContainerDied","Data":"cdd9e5087dbe100b4200dc045cc6536b7bd5644b604c1ad48cd724f12116a2d5"} Jan 29 16:26:32 crc kubenswrapper[4714]: I0129 16:26:32.074990 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg" event={"ID":"dd598a7a-34ba-4392-908b-c18d89648bb5","Type":"ContainerStarted","Data":"2c5ce017fddcc26d58ff7fc5c31e1e866cf3cb41da47bd4a24dc52b2e88e55e7"} Jan 29 16:26:35 crc kubenswrapper[4714]: I0129 16:26:35.093636 4714 generic.go:334] "Generic (PLEG): container finished" podID="dd598a7a-34ba-4392-908b-c18d89648bb5" containerID="a293414bfe6fecbd34f2097ce525abe6d50f764495aa3e4545c1f2cdb4d889ff" exitCode=0 Jan 29 16:26:35 crc kubenswrapper[4714]: I0129 16:26:35.093730 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg" event={"ID":"dd598a7a-34ba-4392-908b-c18d89648bb5","Type":"ContainerDied","Data":"a293414bfe6fecbd34f2097ce525abe6d50f764495aa3e4545c1f2cdb4d889ff"} Jan 29 16:26:35 crc kubenswrapper[4714]: E0129 16:26:35.309637 4714 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 16:26:35 crc kubenswrapper[4714]: E0129 16:26:35.310108 4714 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rn67x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-w58tg_openshift-marketplace(db1a5ea4-712c-4ec8-aeae-c44998fa51e1): ErrImagePull: initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 16:26:35 crc kubenswrapper[4714]: E0129 16:26:35.311305 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/certified-operators-w58tg" podUID="db1a5ea4-712c-4ec8-aeae-c44998fa51e1" Jan 29 16:26:36 crc kubenswrapper[4714]: E0129 16:26:36.311050 4714 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 16:26:36 crc kubenswrapper[4714]: E0129 16:26:36.311237 4714 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tmm88,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-wdwq5_openshift-marketplace(8c12ad14-f878-42a1-a168-bad4026ec2dd): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 16:26:36 crc kubenswrapper[4714]: E0129 16:26:36.312761 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:26:37 crc kubenswrapper[4714]: I0129 16:26:37.111524 4714 generic.go:334] "Generic (PLEG): container finished" podID="dd598a7a-34ba-4392-908b-c18d89648bb5" containerID="e9b290e5fae1b9ebd91874f0c7f54baf70c50604b0924e9b333424187e1578aa" exitCode=0 Jan 29 16:26:37 crc kubenswrapper[4714]: I0129 16:26:37.111691 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg" event={"ID":"dd598a7a-34ba-4392-908b-c18d89648bb5","Type":"ContainerDied","Data":"e9b290e5fae1b9ebd91874f0c7f54baf70c50604b0924e9b333424187e1578aa"} Jan 29 16:26:38 crc kubenswrapper[4714]: I0129 16:26:38.445273 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg" Jan 29 16:26:38 crc kubenswrapper[4714]: I0129 16:26:38.498257 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlxp5\" (UniqueName: \"kubernetes.io/projected/dd598a7a-34ba-4392-908b-c18d89648bb5-kube-api-access-wlxp5\") pod \"dd598a7a-34ba-4392-908b-c18d89648bb5\" (UID: \"dd598a7a-34ba-4392-908b-c18d89648bb5\") " Jan 29 16:26:38 crc kubenswrapper[4714]: I0129 16:26:38.498328 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dd598a7a-34ba-4392-908b-c18d89648bb5-util\") pod \"dd598a7a-34ba-4392-908b-c18d89648bb5\" (UID: \"dd598a7a-34ba-4392-908b-c18d89648bb5\") " Jan 29 16:26:38 crc kubenswrapper[4714]: I0129 16:26:38.498376 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dd598a7a-34ba-4392-908b-c18d89648bb5-bundle\") pod \"dd598a7a-34ba-4392-908b-c18d89648bb5\" (UID: \"dd598a7a-34ba-4392-908b-c18d89648bb5\") " Jan 29 16:26:38 crc kubenswrapper[4714]: I0129 16:26:38.499511 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd598a7a-34ba-4392-908b-c18d89648bb5-bundle" (OuterVolumeSpecName: "bundle") pod "dd598a7a-34ba-4392-908b-c18d89648bb5" (UID: "dd598a7a-34ba-4392-908b-c18d89648bb5"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:26:38 crc kubenswrapper[4714]: I0129 16:26:38.503779 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd598a7a-34ba-4392-908b-c18d89648bb5-kube-api-access-wlxp5" (OuterVolumeSpecName: "kube-api-access-wlxp5") pod "dd598a7a-34ba-4392-908b-c18d89648bb5" (UID: "dd598a7a-34ba-4392-908b-c18d89648bb5"). InnerVolumeSpecName "kube-api-access-wlxp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:26:38 crc kubenswrapper[4714]: I0129 16:26:38.509094 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd598a7a-34ba-4392-908b-c18d89648bb5-util" (OuterVolumeSpecName: "util") pod "dd598a7a-34ba-4392-908b-c18d89648bb5" (UID: "dd598a7a-34ba-4392-908b-c18d89648bb5"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:26:38 crc kubenswrapper[4714]: I0129 16:26:38.600453 4714 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dd598a7a-34ba-4392-908b-c18d89648bb5-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 16:26:38 crc kubenswrapper[4714]: I0129 16:26:38.600501 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlxp5\" (UniqueName: \"kubernetes.io/projected/dd598a7a-34ba-4392-908b-c18d89648bb5-kube-api-access-wlxp5\") on node \"crc\" DevicePath \"\"" Jan 29 16:26:38 crc kubenswrapper[4714]: I0129 16:26:38.600513 4714 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dd598a7a-34ba-4392-908b-c18d89648bb5-util\") on node \"crc\" DevicePath \"\"" Jan 29 16:26:39 crc kubenswrapper[4714]: I0129 16:26:39.135308 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg" event={"ID":"dd598a7a-34ba-4392-908b-c18d89648bb5","Type":"ContainerDied","Data":"2c5ce017fddcc26d58ff7fc5c31e1e866cf3cb41da47bd4a24dc52b2e88e55e7"} Jan 29 16:26:39 crc kubenswrapper[4714]: I0129 16:26:39.135393 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c5ce017fddcc26d58ff7fc5c31e1e866cf3cb41da47bd4a24dc52b2e88e55e7" Jan 29 16:26:39 crc kubenswrapper[4714]: I0129 16:26:39.135522 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg" Jan 29 16:26:49 crc kubenswrapper[4714]: E0129 16:26:49.188413 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-w58tg" podUID="db1a5ea4-712c-4ec8-aeae-c44998fa51e1" Jan 29 16:26:49 crc kubenswrapper[4714]: E0129 16:26:49.188413 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:26:49 crc kubenswrapper[4714]: I0129 16:26:49.349326 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28"] Jan 29 16:26:49 crc kubenswrapper[4714]: E0129 16:26:49.349558 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd598a7a-34ba-4392-908b-c18d89648bb5" containerName="util" Jan 29 16:26:49 crc kubenswrapper[4714]: I0129 16:26:49.349570 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd598a7a-34ba-4392-908b-c18d89648bb5" containerName="util" Jan 29 16:26:49 crc kubenswrapper[4714]: E0129 16:26:49.349584 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd598a7a-34ba-4392-908b-c18d89648bb5" containerName="pull" Jan 29 16:26:49 crc kubenswrapper[4714]: I0129 16:26:49.349589 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd598a7a-34ba-4392-908b-c18d89648bb5" containerName="pull" Jan 29 16:26:49 crc kubenswrapper[4714]: E0129 16:26:49.349601 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd598a7a-34ba-4392-908b-c18d89648bb5" containerName="extract" Jan 29 16:26:49 crc kubenswrapper[4714]: I0129 16:26:49.349607 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd598a7a-34ba-4392-908b-c18d89648bb5" containerName="extract" Jan 29 16:26:49 crc kubenswrapper[4714]: I0129 16:26:49.349703 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd598a7a-34ba-4392-908b-c18d89648bb5" containerName="extract" Jan 29 16:26:49 crc kubenswrapper[4714]: I0129 16:26:49.350105 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" Jan 29 16:26:49 crc kubenswrapper[4714]: I0129 16:26:49.351826 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-m5648" Jan 29 16:26:49 crc kubenswrapper[4714]: I0129 16:26:49.351857 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Jan 29 16:26:49 crc kubenswrapper[4714]: I0129 16:26:49.364061 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28"] Jan 29 16:26:49 crc kubenswrapper[4714]: I0129 16:26:49.499104 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5d602ee5-4171-4dc7-9852-88c6019696e1-apiservice-cert\") pod \"keystone-operator-controller-manager-5b97656f4c-wwx28\" (UID: \"5d602ee5-4171-4dc7-9852-88c6019696e1\") " pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" Jan 29 16:26:49 crc kubenswrapper[4714]: I0129 16:26:49.499177 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5d602ee5-4171-4dc7-9852-88c6019696e1-webhook-cert\") pod \"keystone-operator-controller-manager-5b97656f4c-wwx28\" (UID: \"5d602ee5-4171-4dc7-9852-88c6019696e1\") " pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" Jan 29 16:26:49 crc kubenswrapper[4714]: I0129 16:26:49.499442 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7sdp\" (UniqueName: \"kubernetes.io/projected/5d602ee5-4171-4dc7-9852-88c6019696e1-kube-api-access-p7sdp\") pod \"keystone-operator-controller-manager-5b97656f4c-wwx28\" (UID: \"5d602ee5-4171-4dc7-9852-88c6019696e1\") " pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" Jan 29 16:26:49 crc kubenswrapper[4714]: I0129 16:26:49.600584 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5d602ee5-4171-4dc7-9852-88c6019696e1-webhook-cert\") pod \"keystone-operator-controller-manager-5b97656f4c-wwx28\" (UID: \"5d602ee5-4171-4dc7-9852-88c6019696e1\") " pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" Jan 29 16:26:49 crc kubenswrapper[4714]: I0129 16:26:49.600674 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7sdp\" (UniqueName: \"kubernetes.io/projected/5d602ee5-4171-4dc7-9852-88c6019696e1-kube-api-access-p7sdp\") pod \"keystone-operator-controller-manager-5b97656f4c-wwx28\" (UID: \"5d602ee5-4171-4dc7-9852-88c6019696e1\") " pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" Jan 29 16:26:49 crc kubenswrapper[4714]: I0129 16:26:49.600725 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5d602ee5-4171-4dc7-9852-88c6019696e1-apiservice-cert\") pod \"keystone-operator-controller-manager-5b97656f4c-wwx28\" (UID: \"5d602ee5-4171-4dc7-9852-88c6019696e1\") " pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" Jan 29 16:26:49 crc kubenswrapper[4714]: I0129 16:26:49.607761 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5d602ee5-4171-4dc7-9852-88c6019696e1-webhook-cert\") pod \"keystone-operator-controller-manager-5b97656f4c-wwx28\" (UID: \"5d602ee5-4171-4dc7-9852-88c6019696e1\") " pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" Jan 29 16:26:49 crc kubenswrapper[4714]: I0129 16:26:49.609762 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5d602ee5-4171-4dc7-9852-88c6019696e1-apiservice-cert\") pod \"keystone-operator-controller-manager-5b97656f4c-wwx28\" (UID: \"5d602ee5-4171-4dc7-9852-88c6019696e1\") " pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" Jan 29 16:26:49 crc kubenswrapper[4714]: I0129 16:26:49.621981 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7sdp\" (UniqueName: \"kubernetes.io/projected/5d602ee5-4171-4dc7-9852-88c6019696e1-kube-api-access-p7sdp\") pod \"keystone-operator-controller-manager-5b97656f4c-wwx28\" (UID: \"5d602ee5-4171-4dc7-9852-88c6019696e1\") " pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" Jan 29 16:26:49 crc kubenswrapper[4714]: I0129 16:26:49.666593 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" Jan 29 16:26:49 crc kubenswrapper[4714]: I0129 16:26:49.918003 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28"] Jan 29 16:26:49 crc kubenswrapper[4714]: W0129 16:26:49.924352 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d602ee5_4171_4dc7_9852_88c6019696e1.slice/crio-68e4148d363b3ff81741fb86667cb2d686f613c7db6f1f53e639c87301659f49 WatchSource:0}: Error finding container 68e4148d363b3ff81741fb86667cb2d686f613c7db6f1f53e639c87301659f49: Status 404 returned error can't find the container with id 68e4148d363b3ff81741fb86667cb2d686f613c7db6f1f53e639c87301659f49 Jan 29 16:26:50 crc kubenswrapper[4714]: I0129 16:26:50.219897 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" event={"ID":"5d602ee5-4171-4dc7-9852-88c6019696e1","Type":"ContainerStarted","Data":"68e4148d363b3ff81741fb86667cb2d686f613c7db6f1f53e639c87301659f49"} Jan 29 16:26:55 crc kubenswrapper[4714]: I0129 16:26:55.262686 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" event={"ID":"5d602ee5-4171-4dc7-9852-88c6019696e1","Type":"ContainerStarted","Data":"0389615fc2e1a33814fa46fa1999a6f01822d1c7cd6bb84e89e87e7102d3acc5"} Jan 29 16:26:55 crc kubenswrapper[4714]: I0129 16:26:55.263207 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" Jan 29 16:26:55 crc kubenswrapper[4714]: I0129 16:26:55.286529 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" podStartSLOduration=1.760099205 podStartE2EDuration="6.28650597s" podCreationTimestamp="2026-01-29 16:26:49 +0000 UTC" firstStartedPulling="2026-01-29 16:26:49.927044747 +0000 UTC m=+1016.447545867" lastFinishedPulling="2026-01-29 16:26:54.453451522 +0000 UTC m=+1020.973952632" observedRunningTime="2026-01-29 16:26:55.27901753 +0000 UTC m=+1021.799518670" watchObservedRunningTime="2026-01-29 16:26:55.28650597 +0000 UTC m=+1021.807007090" Jan 29 16:26:57 crc kubenswrapper[4714]: I0129 16:26:57.844550 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:26:57 crc kubenswrapper[4714]: I0129 16:26:57.845577 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:26:58 crc kubenswrapper[4714]: I0129 16:26:58.296571 4714 generic.go:334] "Generic (PLEG): container finished" podID="55e23ac1-a89b-4689-a17d-bee875f7783e" containerID="eb95fbc4965d0aeffea6fce3a32742ee3a1bc5446fd33b5341ed6d4be042d493" exitCode=0 Jan 29 16:26:58 crc kubenswrapper[4714]: I0129 16:26:58.296638 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"55e23ac1-a89b-4689-a17d-bee875f7783e","Type":"ContainerDied","Data":"eb95fbc4965d0aeffea6fce3a32742ee3a1bc5446fd33b5341ed6d4be042d493"} Jan 29 16:26:59 crc kubenswrapper[4714]: I0129 16:26:59.306481 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"55e23ac1-a89b-4689-a17d-bee875f7783e","Type":"ContainerStarted","Data":"ee125ef7ce148229ad82b22d2a7578dd54b32a1f3c67998b7176056878a4676a"} Jan 29 16:26:59 crc kubenswrapper[4714]: I0129 16:26:59.306839 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:26:59 crc kubenswrapper[4714]: I0129 16:26:59.328794 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/rabbitmq-server-0" podStartSLOduration=36.688712317 podStartE2EDuration="45.328770933s" podCreationTimestamp="2026-01-29 16:26:14 +0000 UTC" firstStartedPulling="2026-01-29 16:26:16.523093735 +0000 UTC m=+983.043594855" lastFinishedPulling="2026-01-29 16:26:25.163152341 +0000 UTC m=+991.683653471" observedRunningTime="2026-01-29 16:26:59.327693223 +0000 UTC m=+1025.848194343" watchObservedRunningTime="2026-01-29 16:26:59.328770933 +0000 UTC m=+1025.849272073" Jan 29 16:26:59 crc kubenswrapper[4714]: I0129 16:26:59.672017 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" Jan 29 16:27:01 crc kubenswrapper[4714]: E0129 16:27:01.187626 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:27:02 crc kubenswrapper[4714]: I0129 16:27:02.340753 4714 generic.go:334] "Generic (PLEG): container finished" podID="db1a5ea4-712c-4ec8-aeae-c44998fa51e1" containerID="4b864e9c0bb2f4364a8efa096fde9bcf1de20bf04e06b58bcc7485e4462f93c5" exitCode=0 Jan 29 16:27:02 crc kubenswrapper[4714]: I0129 16:27:02.340830 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w58tg" event={"ID":"db1a5ea4-712c-4ec8-aeae-c44998fa51e1","Type":"ContainerDied","Data":"4b864e9c0bb2f4364a8efa096fde9bcf1de20bf04e06b58bcc7485e4462f93c5"} Jan 29 16:27:03 crc kubenswrapper[4714]: I0129 16:27:03.348011 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w58tg" event={"ID":"db1a5ea4-712c-4ec8-aeae-c44998fa51e1","Type":"ContainerStarted","Data":"87939f973c741f22ef30a3e11e1a7e3cde2e6dfae61483c34dee22312eab0f29"} Jan 29 16:27:03 crc kubenswrapper[4714]: I0129 16:27:03.362537 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-w58tg" podStartSLOduration=2.63832229 podStartE2EDuration="45.362516708s" podCreationTimestamp="2026-01-29 16:26:18 +0000 UTC" firstStartedPulling="2026-01-29 16:26:20.012068362 +0000 UTC m=+986.532569482" lastFinishedPulling="2026-01-29 16:27:02.73626278 +0000 UTC m=+1029.256763900" observedRunningTime="2026-01-29 16:27:03.361268913 +0000 UTC m=+1029.881770043" watchObservedRunningTime="2026-01-29 16:27:03.362516708 +0000 UTC m=+1029.883017828" Jan 29 16:27:09 crc kubenswrapper[4714]: I0129 16:27:09.134202 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-w58tg" Jan 29 16:27:09 crc kubenswrapper[4714]: I0129 16:27:09.135374 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-w58tg" Jan 29 16:27:09 crc kubenswrapper[4714]: I0129 16:27:09.191066 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-index-d7f6m"] Jan 29 16:27:09 crc kubenswrapper[4714]: I0129 16:27:09.205266 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-index-d7f6m" Jan 29 16:27:09 crc kubenswrapper[4714]: I0129 16:27:09.210722 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-index-dockercfg-tgxd8" Jan 29 16:27:09 crc kubenswrapper[4714]: I0129 16:27:09.216143 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-index-d7f6m"] Jan 29 16:27:09 crc kubenswrapper[4714]: I0129 16:27:09.222714 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-w58tg" Jan 29 16:27:09 crc kubenswrapper[4714]: I0129 16:27:09.289031 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6lgj\" (UniqueName: \"kubernetes.io/projected/87506df3-b56a-4598-8309-e865dc93cf53-kube-api-access-m6lgj\") pod \"cinder-operator-index-d7f6m\" (UID: \"87506df3-b56a-4598-8309-e865dc93cf53\") " pod="openstack-operators/cinder-operator-index-d7f6m" Jan 29 16:27:09 crc kubenswrapper[4714]: I0129 16:27:09.392426 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6lgj\" (UniqueName: \"kubernetes.io/projected/87506df3-b56a-4598-8309-e865dc93cf53-kube-api-access-m6lgj\") pod \"cinder-operator-index-d7f6m\" (UID: \"87506df3-b56a-4598-8309-e865dc93cf53\") " pod="openstack-operators/cinder-operator-index-d7f6m" Jan 29 16:27:09 crc kubenswrapper[4714]: I0129 16:27:09.419738 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6lgj\" (UniqueName: \"kubernetes.io/projected/87506df3-b56a-4598-8309-e865dc93cf53-kube-api-access-m6lgj\") pod \"cinder-operator-index-d7f6m\" (UID: \"87506df3-b56a-4598-8309-e865dc93cf53\") " pod="openstack-operators/cinder-operator-index-d7f6m" Jan 29 16:27:09 crc kubenswrapper[4714]: I0129 16:27:09.448248 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-w58tg" Jan 29 16:27:09 crc kubenswrapper[4714]: I0129 16:27:09.536631 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-index-d7f6m" Jan 29 16:27:09 crc kubenswrapper[4714]: I0129 16:27:09.952268 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-index-d7f6m"] Jan 29 16:27:10 crc kubenswrapper[4714]: I0129 16:27:10.396030 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-index-d7f6m" event={"ID":"87506df3-b56a-4598-8309-e865dc93cf53","Type":"ContainerStarted","Data":"6f87b469ba7b044e8c1285048b5bce1bfd385e6eeae34b8f824127e313741cf2"} Jan 29 16:27:12 crc kubenswrapper[4714]: E0129 16:27:12.194371 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:27:13 crc kubenswrapper[4714]: I0129 16:27:13.427156 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-index-d7f6m" event={"ID":"87506df3-b56a-4598-8309-e865dc93cf53","Type":"ContainerStarted","Data":"ddec768853957c5a9f807fefab5f4e26d2050966765bc9353326a1334ad4a549"} Jan 29 16:27:13 crc kubenswrapper[4714]: I0129 16:27:13.446227 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-index-d7f6m" podStartSLOduration=1.628648934 podStartE2EDuration="4.446193751s" podCreationTimestamp="2026-01-29 16:27:09 +0000 UTC" firstStartedPulling="2026-01-29 16:27:09.967845933 +0000 UTC m=+1036.488347053" lastFinishedPulling="2026-01-29 16:27:12.78539075 +0000 UTC m=+1039.305891870" observedRunningTime="2026-01-29 16:27:13.442675395 +0000 UTC m=+1039.963176535" watchObservedRunningTime="2026-01-29 16:27:13.446193751 +0000 UTC m=+1039.966694871" Jan 29 16:27:13 crc kubenswrapper[4714]: I0129 16:27:13.744055 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w58tg"] Jan 29 16:27:13 crc kubenswrapper[4714]: I0129 16:27:13.744304 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-w58tg" podUID="db1a5ea4-712c-4ec8-aeae-c44998fa51e1" containerName="registry-server" containerID="cri-o://87939f973c741f22ef30a3e11e1a7e3cde2e6dfae61483c34dee22312eab0f29" gracePeriod=2 Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.196191 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w58tg" Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.276423 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db1a5ea4-712c-4ec8-aeae-c44998fa51e1-catalog-content\") pod \"db1a5ea4-712c-4ec8-aeae-c44998fa51e1\" (UID: \"db1a5ea4-712c-4ec8-aeae-c44998fa51e1\") " Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.276503 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rn67x\" (UniqueName: \"kubernetes.io/projected/db1a5ea4-712c-4ec8-aeae-c44998fa51e1-kube-api-access-rn67x\") pod \"db1a5ea4-712c-4ec8-aeae-c44998fa51e1\" (UID: \"db1a5ea4-712c-4ec8-aeae-c44998fa51e1\") " Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.276562 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db1a5ea4-712c-4ec8-aeae-c44998fa51e1-utilities\") pod \"db1a5ea4-712c-4ec8-aeae-c44998fa51e1\" (UID: \"db1a5ea4-712c-4ec8-aeae-c44998fa51e1\") " Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.277506 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db1a5ea4-712c-4ec8-aeae-c44998fa51e1-utilities" (OuterVolumeSpecName: "utilities") pod "db1a5ea4-712c-4ec8-aeae-c44998fa51e1" (UID: "db1a5ea4-712c-4ec8-aeae-c44998fa51e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.297097 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db1a5ea4-712c-4ec8-aeae-c44998fa51e1-kube-api-access-rn67x" (OuterVolumeSpecName: "kube-api-access-rn67x") pod "db1a5ea4-712c-4ec8-aeae-c44998fa51e1" (UID: "db1a5ea4-712c-4ec8-aeae-c44998fa51e1"). InnerVolumeSpecName "kube-api-access-rn67x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.328777 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db1a5ea4-712c-4ec8-aeae-c44998fa51e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db1a5ea4-712c-4ec8-aeae-c44998fa51e1" (UID: "db1a5ea4-712c-4ec8-aeae-c44998fa51e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.378137 4714 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db1a5ea4-712c-4ec8-aeae-c44998fa51e1-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.378176 4714 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db1a5ea4-712c-4ec8-aeae-c44998fa51e1-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.378193 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rn67x\" (UniqueName: \"kubernetes.io/projected/db1a5ea4-712c-4ec8-aeae-c44998fa51e1-kube-api-access-rn67x\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.435313 4714 generic.go:334] "Generic (PLEG): container finished" podID="db1a5ea4-712c-4ec8-aeae-c44998fa51e1" containerID="87939f973c741f22ef30a3e11e1a7e3cde2e6dfae61483c34dee22312eab0f29" exitCode=0 Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.435381 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w58tg" event={"ID":"db1a5ea4-712c-4ec8-aeae-c44998fa51e1","Type":"ContainerDied","Data":"87939f973c741f22ef30a3e11e1a7e3cde2e6dfae61483c34dee22312eab0f29"} Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.435434 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w58tg" event={"ID":"db1a5ea4-712c-4ec8-aeae-c44998fa51e1","Type":"ContainerDied","Data":"fa3e0fa56a351fb40b73e185e0120625630ac7fd4a18cf99fdfb008b5ae7fadf"} Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.435435 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w58tg" Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.435454 4714 scope.go:117] "RemoveContainer" containerID="87939f973c741f22ef30a3e11e1a7e3cde2e6dfae61483c34dee22312eab0f29" Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.453242 4714 scope.go:117] "RemoveContainer" containerID="4b864e9c0bb2f4364a8efa096fde9bcf1de20bf04e06b58bcc7485e4462f93c5" Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.461808 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w58tg"] Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.468155 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-w58tg"] Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.529752 4714 scope.go:117] "RemoveContainer" containerID="977047a32f415a72ab936afe2d1774146105933da01176d2aabeb71559b4afad" Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.545549 4714 scope.go:117] "RemoveContainer" containerID="87939f973c741f22ef30a3e11e1a7e3cde2e6dfae61483c34dee22312eab0f29" Jan 29 16:27:14 crc kubenswrapper[4714]: E0129 16:27:14.546033 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87939f973c741f22ef30a3e11e1a7e3cde2e6dfae61483c34dee22312eab0f29\": container with ID starting with 87939f973c741f22ef30a3e11e1a7e3cde2e6dfae61483c34dee22312eab0f29 not found: ID does not exist" containerID="87939f973c741f22ef30a3e11e1a7e3cde2e6dfae61483c34dee22312eab0f29" Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.546084 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87939f973c741f22ef30a3e11e1a7e3cde2e6dfae61483c34dee22312eab0f29"} err="failed to get container status \"87939f973c741f22ef30a3e11e1a7e3cde2e6dfae61483c34dee22312eab0f29\": rpc error: code = NotFound desc = could not find container \"87939f973c741f22ef30a3e11e1a7e3cde2e6dfae61483c34dee22312eab0f29\": container with ID starting with 87939f973c741f22ef30a3e11e1a7e3cde2e6dfae61483c34dee22312eab0f29 not found: ID does not exist" Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.546110 4714 scope.go:117] "RemoveContainer" containerID="4b864e9c0bb2f4364a8efa096fde9bcf1de20bf04e06b58bcc7485e4462f93c5" Jan 29 16:27:14 crc kubenswrapper[4714]: E0129 16:27:14.546664 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b864e9c0bb2f4364a8efa096fde9bcf1de20bf04e06b58bcc7485e4462f93c5\": container with ID starting with 4b864e9c0bb2f4364a8efa096fde9bcf1de20bf04e06b58bcc7485e4462f93c5 not found: ID does not exist" containerID="4b864e9c0bb2f4364a8efa096fde9bcf1de20bf04e06b58bcc7485e4462f93c5" Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.546694 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b864e9c0bb2f4364a8efa096fde9bcf1de20bf04e06b58bcc7485e4462f93c5"} err="failed to get container status \"4b864e9c0bb2f4364a8efa096fde9bcf1de20bf04e06b58bcc7485e4462f93c5\": rpc error: code = NotFound desc = could not find container \"4b864e9c0bb2f4364a8efa096fde9bcf1de20bf04e06b58bcc7485e4462f93c5\": container with ID starting with 4b864e9c0bb2f4364a8efa096fde9bcf1de20bf04e06b58bcc7485e4462f93c5 not found: ID does not exist" Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.546717 4714 scope.go:117] "RemoveContainer" containerID="977047a32f415a72ab936afe2d1774146105933da01176d2aabeb71559b4afad" Jan 29 16:27:14 crc kubenswrapper[4714]: E0129 16:27:14.547062 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"977047a32f415a72ab936afe2d1774146105933da01176d2aabeb71559b4afad\": container with ID starting with 977047a32f415a72ab936afe2d1774146105933da01176d2aabeb71559b4afad not found: ID does not exist" containerID="977047a32f415a72ab936afe2d1774146105933da01176d2aabeb71559b4afad" Jan 29 16:27:14 crc kubenswrapper[4714]: I0129 16:27:14.547080 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"977047a32f415a72ab936afe2d1774146105933da01176d2aabeb71559b4afad"} err="failed to get container status \"977047a32f415a72ab936afe2d1774146105933da01176d2aabeb71559b4afad\": rpc error: code = NotFound desc = could not find container \"977047a32f415a72ab936afe2d1774146105933da01176d2aabeb71559b4afad\": container with ID starting with 977047a32f415a72ab936afe2d1774146105933da01176d2aabeb71559b4afad not found: ID does not exist" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.362908 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-db-create-4hhfq"] Jan 29 16:27:15 crc kubenswrapper[4714]: E0129 16:27:15.363427 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db1a5ea4-712c-4ec8-aeae-c44998fa51e1" containerName="registry-server" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.363504 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="db1a5ea4-712c-4ec8-aeae-c44998fa51e1" containerName="registry-server" Jan 29 16:27:15 crc kubenswrapper[4714]: E0129 16:27:15.363587 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db1a5ea4-712c-4ec8-aeae-c44998fa51e1" containerName="extract-content" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.363643 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="db1a5ea4-712c-4ec8-aeae-c44998fa51e1" containerName="extract-content" Jan 29 16:27:15 crc kubenswrapper[4714]: E0129 16:27:15.363698 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db1a5ea4-712c-4ec8-aeae-c44998fa51e1" containerName="extract-utilities" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.363752 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="db1a5ea4-712c-4ec8-aeae-c44998fa51e1" containerName="extract-utilities" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.363913 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="db1a5ea4-712c-4ec8-aeae-c44998fa51e1" containerName="registry-server" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.364431 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-create-4hhfq" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.376741 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db-create-4hhfq"] Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.383084 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-52f2-account-create-update-5x6mg"] Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.383915 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-52f2-account-create-update-5x6mg" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.387045 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-db-secret" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.407142 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-52f2-account-create-update-5x6mg"] Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.494003 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6c31118-0f0d-46fb-a9fc-d135e234fe41-operator-scripts\") pod \"keystone-52f2-account-create-update-5x6mg\" (UID: \"f6c31118-0f0d-46fb-a9fc-d135e234fe41\") " pod="cinder-kuttl-tests/keystone-52f2-account-create-update-5x6mg" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.494049 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znwpf\" (UniqueName: \"kubernetes.io/projected/f6c31118-0f0d-46fb-a9fc-d135e234fe41-kube-api-access-znwpf\") pod \"keystone-52f2-account-create-update-5x6mg\" (UID: \"f6c31118-0f0d-46fb-a9fc-d135e234fe41\") " pod="cinder-kuttl-tests/keystone-52f2-account-create-update-5x6mg" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.494081 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bf4c895-a323-452d-8329-cb69a752341c-operator-scripts\") pod \"keystone-db-create-4hhfq\" (UID: \"4bf4c895-a323-452d-8329-cb69a752341c\") " pod="cinder-kuttl-tests/keystone-db-create-4hhfq" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.494113 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq285\" (UniqueName: \"kubernetes.io/projected/4bf4c895-a323-452d-8329-cb69a752341c-kube-api-access-sq285\") pod \"keystone-db-create-4hhfq\" (UID: \"4bf4c895-a323-452d-8329-cb69a752341c\") " pod="cinder-kuttl-tests/keystone-db-create-4hhfq" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.595088 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6c31118-0f0d-46fb-a9fc-d135e234fe41-operator-scripts\") pod \"keystone-52f2-account-create-update-5x6mg\" (UID: \"f6c31118-0f0d-46fb-a9fc-d135e234fe41\") " pod="cinder-kuttl-tests/keystone-52f2-account-create-update-5x6mg" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.595143 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znwpf\" (UniqueName: \"kubernetes.io/projected/f6c31118-0f0d-46fb-a9fc-d135e234fe41-kube-api-access-znwpf\") pod \"keystone-52f2-account-create-update-5x6mg\" (UID: \"f6c31118-0f0d-46fb-a9fc-d135e234fe41\") " pod="cinder-kuttl-tests/keystone-52f2-account-create-update-5x6mg" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.595180 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bf4c895-a323-452d-8329-cb69a752341c-operator-scripts\") pod \"keystone-db-create-4hhfq\" (UID: \"4bf4c895-a323-452d-8329-cb69a752341c\") " pod="cinder-kuttl-tests/keystone-db-create-4hhfq" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.595204 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq285\" (UniqueName: \"kubernetes.io/projected/4bf4c895-a323-452d-8329-cb69a752341c-kube-api-access-sq285\") pod \"keystone-db-create-4hhfq\" (UID: \"4bf4c895-a323-452d-8329-cb69a752341c\") " pod="cinder-kuttl-tests/keystone-db-create-4hhfq" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.596018 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6c31118-0f0d-46fb-a9fc-d135e234fe41-operator-scripts\") pod \"keystone-52f2-account-create-update-5x6mg\" (UID: \"f6c31118-0f0d-46fb-a9fc-d135e234fe41\") " pod="cinder-kuttl-tests/keystone-52f2-account-create-update-5x6mg" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.596069 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bf4c895-a323-452d-8329-cb69a752341c-operator-scripts\") pod \"keystone-db-create-4hhfq\" (UID: \"4bf4c895-a323-452d-8329-cb69a752341c\") " pod="cinder-kuttl-tests/keystone-db-create-4hhfq" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.613648 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znwpf\" (UniqueName: \"kubernetes.io/projected/f6c31118-0f0d-46fb-a9fc-d135e234fe41-kube-api-access-znwpf\") pod \"keystone-52f2-account-create-update-5x6mg\" (UID: \"f6c31118-0f0d-46fb-a9fc-d135e234fe41\") " pod="cinder-kuttl-tests/keystone-52f2-account-create-update-5x6mg" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.615050 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq285\" (UniqueName: \"kubernetes.io/projected/4bf4c895-a323-452d-8329-cb69a752341c-kube-api-access-sq285\") pod \"keystone-db-create-4hhfq\" (UID: \"4bf4c895-a323-452d-8329-cb69a752341c\") " pod="cinder-kuttl-tests/keystone-db-create-4hhfq" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.677284 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-create-4hhfq" Jan 29 16:27:15 crc kubenswrapper[4714]: I0129 16:27:15.696697 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-52f2-account-create-update-5x6mg" Jan 29 16:27:16 crc kubenswrapper[4714]: I0129 16:27:16.093688 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:27:16 crc kubenswrapper[4714]: I0129 16:27:16.162759 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db-create-4hhfq"] Jan 29 16:27:16 crc kubenswrapper[4714]: I0129 16:27:16.194570 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db1a5ea4-712c-4ec8-aeae-c44998fa51e1" path="/var/lib/kubelet/pods/db1a5ea4-712c-4ec8-aeae-c44998fa51e1/volumes" Jan 29 16:27:16 crc kubenswrapper[4714]: I0129 16:27:16.233720 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-52f2-account-create-update-5x6mg"] Jan 29 16:27:16 crc kubenswrapper[4714]: I0129 16:27:16.451275 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-create-4hhfq" event={"ID":"4bf4c895-a323-452d-8329-cb69a752341c","Type":"ContainerStarted","Data":"958fc7292af56ecfa7d5c5a7066233d1295c2b0c82a6e8e4646901914aabf005"} Jan 29 16:27:16 crc kubenswrapper[4714]: I0129 16:27:16.451364 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-create-4hhfq" event={"ID":"4bf4c895-a323-452d-8329-cb69a752341c","Type":"ContainerStarted","Data":"143876cbf9d3526f84e82f093d097878163e840a2b90cc081656f409d4bac93e"} Jan 29 16:27:16 crc kubenswrapper[4714]: I0129 16:27:16.454280 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-52f2-account-create-update-5x6mg" event={"ID":"f6c31118-0f0d-46fb-a9fc-d135e234fe41","Type":"ContainerStarted","Data":"57df90298e52fbd874f183f4f349569b890f49a89aba1583825eacf13909d613"} Jan 29 16:27:16 crc kubenswrapper[4714]: I0129 16:27:16.454319 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-52f2-account-create-update-5x6mg" event={"ID":"f6c31118-0f0d-46fb-a9fc-d135e234fe41","Type":"ContainerStarted","Data":"978ec753876dcd68b52989b0c105fe5d3cde2a954d0e4d72d8a1ef936aa61c3d"} Jan 29 16:27:16 crc kubenswrapper[4714]: I0129 16:27:16.502762 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone-db-create-4hhfq" podStartSLOduration=1.502743765 podStartE2EDuration="1.502743765s" podCreationTimestamp="2026-01-29 16:27:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:27:16.479457255 +0000 UTC m=+1042.999958375" watchObservedRunningTime="2026-01-29 16:27:16.502743765 +0000 UTC m=+1043.023244885" Jan 29 16:27:17 crc kubenswrapper[4714]: I0129 16:27:17.464221 4714 generic.go:334] "Generic (PLEG): container finished" podID="f6c31118-0f0d-46fb-a9fc-d135e234fe41" containerID="57df90298e52fbd874f183f4f349569b890f49a89aba1583825eacf13909d613" exitCode=0 Jan 29 16:27:17 crc kubenswrapper[4714]: I0129 16:27:17.464284 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-52f2-account-create-update-5x6mg" event={"ID":"f6c31118-0f0d-46fb-a9fc-d135e234fe41","Type":"ContainerDied","Data":"57df90298e52fbd874f183f4f349569b890f49a89aba1583825eacf13909d613"} Jan 29 16:27:17 crc kubenswrapper[4714]: I0129 16:27:17.466776 4714 generic.go:334] "Generic (PLEG): container finished" podID="4bf4c895-a323-452d-8329-cb69a752341c" containerID="958fc7292af56ecfa7d5c5a7066233d1295c2b0c82a6e8e4646901914aabf005" exitCode=0 Jan 29 16:27:17 crc kubenswrapper[4714]: I0129 16:27:17.466813 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-create-4hhfq" event={"ID":"4bf4c895-a323-452d-8329-cb69a752341c","Type":"ContainerDied","Data":"958fc7292af56ecfa7d5c5a7066233d1295c2b0c82a6e8e4646901914aabf005"} Jan 29 16:27:18 crc kubenswrapper[4714]: I0129 16:27:18.831358 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-create-4hhfq" Jan 29 16:27:18 crc kubenswrapper[4714]: I0129 16:27:18.837536 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-52f2-account-create-update-5x6mg" Jan 29 16:27:18 crc kubenswrapper[4714]: I0129 16:27:18.941176 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sq285\" (UniqueName: \"kubernetes.io/projected/4bf4c895-a323-452d-8329-cb69a752341c-kube-api-access-sq285\") pod \"4bf4c895-a323-452d-8329-cb69a752341c\" (UID: \"4bf4c895-a323-452d-8329-cb69a752341c\") " Jan 29 16:27:18 crc kubenswrapper[4714]: I0129 16:27:18.941334 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bf4c895-a323-452d-8329-cb69a752341c-operator-scripts\") pod \"4bf4c895-a323-452d-8329-cb69a752341c\" (UID: \"4bf4c895-a323-452d-8329-cb69a752341c\") " Jan 29 16:27:18 crc kubenswrapper[4714]: I0129 16:27:18.941376 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6c31118-0f0d-46fb-a9fc-d135e234fe41-operator-scripts\") pod \"f6c31118-0f0d-46fb-a9fc-d135e234fe41\" (UID: \"f6c31118-0f0d-46fb-a9fc-d135e234fe41\") " Jan 29 16:27:18 crc kubenswrapper[4714]: I0129 16:27:18.941432 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znwpf\" (UniqueName: \"kubernetes.io/projected/f6c31118-0f0d-46fb-a9fc-d135e234fe41-kube-api-access-znwpf\") pod \"f6c31118-0f0d-46fb-a9fc-d135e234fe41\" (UID: \"f6c31118-0f0d-46fb-a9fc-d135e234fe41\") " Jan 29 16:27:18 crc kubenswrapper[4714]: I0129 16:27:18.942270 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bf4c895-a323-452d-8329-cb69a752341c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4bf4c895-a323-452d-8329-cb69a752341c" (UID: "4bf4c895-a323-452d-8329-cb69a752341c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:27:18 crc kubenswrapper[4714]: I0129 16:27:18.942337 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6c31118-0f0d-46fb-a9fc-d135e234fe41-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f6c31118-0f0d-46fb-a9fc-d135e234fe41" (UID: "f6c31118-0f0d-46fb-a9fc-d135e234fe41"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:27:18 crc kubenswrapper[4714]: I0129 16:27:18.949076 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bf4c895-a323-452d-8329-cb69a752341c-kube-api-access-sq285" (OuterVolumeSpecName: "kube-api-access-sq285") pod "4bf4c895-a323-452d-8329-cb69a752341c" (UID: "4bf4c895-a323-452d-8329-cb69a752341c"). InnerVolumeSpecName "kube-api-access-sq285". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:27:18 crc kubenswrapper[4714]: I0129 16:27:18.952144 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6c31118-0f0d-46fb-a9fc-d135e234fe41-kube-api-access-znwpf" (OuterVolumeSpecName: "kube-api-access-znwpf") pod "f6c31118-0f0d-46fb-a9fc-d135e234fe41" (UID: "f6c31118-0f0d-46fb-a9fc-d135e234fe41"). InnerVolumeSpecName "kube-api-access-znwpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:27:19 crc kubenswrapper[4714]: I0129 16:27:19.043749 4714 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6c31118-0f0d-46fb-a9fc-d135e234fe41-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:19 crc kubenswrapper[4714]: I0129 16:27:19.043800 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znwpf\" (UniqueName: \"kubernetes.io/projected/f6c31118-0f0d-46fb-a9fc-d135e234fe41-kube-api-access-znwpf\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:19 crc kubenswrapper[4714]: I0129 16:27:19.043823 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sq285\" (UniqueName: \"kubernetes.io/projected/4bf4c895-a323-452d-8329-cb69a752341c-kube-api-access-sq285\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:19 crc kubenswrapper[4714]: I0129 16:27:19.043843 4714 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bf4c895-a323-452d-8329-cb69a752341c-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:19 crc kubenswrapper[4714]: I0129 16:27:19.482409 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-create-4hhfq" event={"ID":"4bf4c895-a323-452d-8329-cb69a752341c","Type":"ContainerDied","Data":"143876cbf9d3526f84e82f093d097878163e840a2b90cc081656f409d4bac93e"} Jan 29 16:27:19 crc kubenswrapper[4714]: I0129 16:27:19.483114 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="143876cbf9d3526f84e82f093d097878163e840a2b90cc081656f409d4bac93e" Jan 29 16:27:19 crc kubenswrapper[4714]: I0129 16:27:19.482474 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-create-4hhfq" Jan 29 16:27:19 crc kubenswrapper[4714]: I0129 16:27:19.484328 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-52f2-account-create-update-5x6mg" event={"ID":"f6c31118-0f0d-46fb-a9fc-d135e234fe41","Type":"ContainerDied","Data":"978ec753876dcd68b52989b0c105fe5d3cde2a954d0e4d72d8a1ef936aa61c3d"} Jan 29 16:27:19 crc kubenswrapper[4714]: I0129 16:27:19.484366 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="978ec753876dcd68b52989b0c105fe5d3cde2a954d0e4d72d8a1ef936aa61c3d" Jan 29 16:27:19 crc kubenswrapper[4714]: I0129 16:27:19.484426 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-52f2-account-create-update-5x6mg" Jan 29 16:27:19 crc kubenswrapper[4714]: I0129 16:27:19.537611 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/cinder-operator-index-d7f6m" Jan 29 16:27:19 crc kubenswrapper[4714]: I0129 16:27:19.537655 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-index-d7f6m" Jan 29 16:27:19 crc kubenswrapper[4714]: I0129 16:27:19.577366 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/cinder-operator-index-d7f6m" Jan 29 16:27:20 crc kubenswrapper[4714]: I0129 16:27:20.532894 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-index-d7f6m" Jan 29 16:27:21 crc kubenswrapper[4714]: I0129 16:27:21.023058 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-6xxc6"] Jan 29 16:27:21 crc kubenswrapper[4714]: E0129 16:27:21.023684 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6c31118-0f0d-46fb-a9fc-d135e234fe41" containerName="mariadb-account-create-update" Jan 29 16:27:21 crc kubenswrapper[4714]: I0129 16:27:21.023708 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6c31118-0f0d-46fb-a9fc-d135e234fe41" containerName="mariadb-account-create-update" Jan 29 16:27:21 crc kubenswrapper[4714]: E0129 16:27:21.023728 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bf4c895-a323-452d-8329-cb69a752341c" containerName="mariadb-database-create" Jan 29 16:27:21 crc kubenswrapper[4714]: I0129 16:27:21.023737 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bf4c895-a323-452d-8329-cb69a752341c" containerName="mariadb-database-create" Jan 29 16:27:21 crc kubenswrapper[4714]: I0129 16:27:21.023890 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bf4c895-a323-452d-8329-cb69a752341c" containerName="mariadb-database-create" Jan 29 16:27:21 crc kubenswrapper[4714]: I0129 16:27:21.023913 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6c31118-0f0d-46fb-a9fc-d135e234fe41" containerName="mariadb-account-create-update" Jan 29 16:27:21 crc kubenswrapper[4714]: I0129 16:27:21.024612 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-sync-6xxc6" Jan 29 16:27:21 crc kubenswrapper[4714]: I0129 16:27:21.026277 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-scripts" Jan 29 16:27:21 crc kubenswrapper[4714]: I0129 16:27:21.026535 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-config-data" Jan 29 16:27:21 crc kubenswrapper[4714]: I0129 16:27:21.027167 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone" Jan 29 16:27:21 crc kubenswrapper[4714]: I0129 16:27:21.032404 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-6xxc6"] Jan 29 16:27:21 crc kubenswrapper[4714]: I0129 16:27:21.032691 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-keystone-dockercfg-6m228" Jan 29 16:27:21 crc kubenswrapper[4714]: I0129 16:27:21.071412 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e33ba3d-9561-441b-b835-fbdb6ce97d23-config-data\") pod \"keystone-db-sync-6xxc6\" (UID: \"6e33ba3d-9561-441b-b835-fbdb6ce97d23\") " pod="cinder-kuttl-tests/keystone-db-sync-6xxc6" Jan 29 16:27:21 crc kubenswrapper[4714]: I0129 16:27:21.071485 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7s7m\" (UniqueName: \"kubernetes.io/projected/6e33ba3d-9561-441b-b835-fbdb6ce97d23-kube-api-access-f7s7m\") pod \"keystone-db-sync-6xxc6\" (UID: \"6e33ba3d-9561-441b-b835-fbdb6ce97d23\") " pod="cinder-kuttl-tests/keystone-db-sync-6xxc6" Jan 29 16:27:21 crc kubenswrapper[4714]: I0129 16:27:21.173359 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7s7m\" (UniqueName: \"kubernetes.io/projected/6e33ba3d-9561-441b-b835-fbdb6ce97d23-kube-api-access-f7s7m\") pod \"keystone-db-sync-6xxc6\" (UID: \"6e33ba3d-9561-441b-b835-fbdb6ce97d23\") " pod="cinder-kuttl-tests/keystone-db-sync-6xxc6" Jan 29 16:27:21 crc kubenswrapper[4714]: I0129 16:27:21.173496 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e33ba3d-9561-441b-b835-fbdb6ce97d23-config-data\") pod \"keystone-db-sync-6xxc6\" (UID: \"6e33ba3d-9561-441b-b835-fbdb6ce97d23\") " pod="cinder-kuttl-tests/keystone-db-sync-6xxc6" Jan 29 16:27:21 crc kubenswrapper[4714]: I0129 16:27:21.179023 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e33ba3d-9561-441b-b835-fbdb6ce97d23-config-data\") pod \"keystone-db-sync-6xxc6\" (UID: \"6e33ba3d-9561-441b-b835-fbdb6ce97d23\") " pod="cinder-kuttl-tests/keystone-db-sync-6xxc6" Jan 29 16:27:21 crc kubenswrapper[4714]: I0129 16:27:21.203782 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7s7m\" (UniqueName: \"kubernetes.io/projected/6e33ba3d-9561-441b-b835-fbdb6ce97d23-kube-api-access-f7s7m\") pod \"keystone-db-sync-6xxc6\" (UID: \"6e33ba3d-9561-441b-b835-fbdb6ce97d23\") " pod="cinder-kuttl-tests/keystone-db-sync-6xxc6" Jan 29 16:27:21 crc kubenswrapper[4714]: I0129 16:27:21.341770 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-sync-6xxc6" Jan 29 16:27:21 crc kubenswrapper[4714]: I0129 16:27:21.831154 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-6xxc6"] Jan 29 16:27:22 crc kubenswrapper[4714]: I0129 16:27:22.406165 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6"] Jan 29 16:27:22 crc kubenswrapper[4714]: I0129 16:27:22.408537 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6" Jan 29 16:27:22 crc kubenswrapper[4714]: I0129 16:27:22.412521 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-hwqbr" Jan 29 16:27:22 crc kubenswrapper[4714]: I0129 16:27:22.419749 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6"] Jan 29 16:27:22 crc kubenswrapper[4714]: I0129 16:27:22.514469 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-sync-6xxc6" event={"ID":"6e33ba3d-9561-441b-b835-fbdb6ce97d23","Type":"ContainerStarted","Data":"cb8e5c568c512500695d36838718c30771a365e70a1f7bf91812aeb8914126a0"} Jan 29 16:27:22 crc kubenswrapper[4714]: I0129 16:27:22.514782 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4wwh\" (UniqueName: \"kubernetes.io/projected/0eecc358-9581-489e-97ae-f600d35a7613-kube-api-access-k4wwh\") pod \"60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6\" (UID: \"0eecc358-9581-489e-97ae-f600d35a7613\") " pod="openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6" Jan 29 16:27:22 crc kubenswrapper[4714]: I0129 16:27:22.514832 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0eecc358-9581-489e-97ae-f600d35a7613-util\") pod \"60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6\" (UID: \"0eecc358-9581-489e-97ae-f600d35a7613\") " pod="openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6" Jan 29 16:27:22 crc kubenswrapper[4714]: I0129 16:27:22.514901 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0eecc358-9581-489e-97ae-f600d35a7613-bundle\") pod \"60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6\" (UID: \"0eecc358-9581-489e-97ae-f600d35a7613\") " pod="openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6" Jan 29 16:27:22 crc kubenswrapper[4714]: I0129 16:27:22.615705 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4wwh\" (UniqueName: \"kubernetes.io/projected/0eecc358-9581-489e-97ae-f600d35a7613-kube-api-access-k4wwh\") pod \"60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6\" (UID: \"0eecc358-9581-489e-97ae-f600d35a7613\") " pod="openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6" Jan 29 16:27:22 crc kubenswrapper[4714]: I0129 16:27:22.615752 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0eecc358-9581-489e-97ae-f600d35a7613-util\") pod \"60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6\" (UID: \"0eecc358-9581-489e-97ae-f600d35a7613\") " pod="openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6" Jan 29 16:27:22 crc kubenswrapper[4714]: I0129 16:27:22.615782 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0eecc358-9581-489e-97ae-f600d35a7613-bundle\") pod \"60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6\" (UID: \"0eecc358-9581-489e-97ae-f600d35a7613\") " pod="openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6" Jan 29 16:27:22 crc kubenswrapper[4714]: I0129 16:27:22.616315 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0eecc358-9581-489e-97ae-f600d35a7613-bundle\") pod \"60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6\" (UID: \"0eecc358-9581-489e-97ae-f600d35a7613\") " pod="openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6" Jan 29 16:27:22 crc kubenswrapper[4714]: I0129 16:27:22.616412 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0eecc358-9581-489e-97ae-f600d35a7613-util\") pod \"60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6\" (UID: \"0eecc358-9581-489e-97ae-f600d35a7613\") " pod="openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6" Jan 29 16:27:22 crc kubenswrapper[4714]: I0129 16:27:22.649614 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4wwh\" (UniqueName: \"kubernetes.io/projected/0eecc358-9581-489e-97ae-f600d35a7613-kube-api-access-k4wwh\") pod \"60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6\" (UID: \"0eecc358-9581-489e-97ae-f600d35a7613\") " pod="openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6" Jan 29 16:27:22 crc kubenswrapper[4714]: I0129 16:27:22.737213 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6" Jan 29 16:27:23 crc kubenswrapper[4714]: I0129 16:27:23.181180 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6"] Jan 29 16:27:23 crc kubenswrapper[4714]: I0129 16:27:23.522668 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6" event={"ID":"0eecc358-9581-489e-97ae-f600d35a7613","Type":"ContainerStarted","Data":"6f5e7b1c14376dcb0e62786390737b48cb04d2a05079d8160df0f9dedc56dfc8"} Jan 29 16:27:23 crc kubenswrapper[4714]: I0129 16:27:23.523217 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6" event={"ID":"0eecc358-9581-489e-97ae-f600d35a7613","Type":"ContainerStarted","Data":"ccf8807833090dda1a00ca4aa3c347ca65828dd70468671dc4659950b5c10bcf"} Jan 29 16:27:24 crc kubenswrapper[4714]: I0129 16:27:24.536217 4714 generic.go:334] "Generic (PLEG): container finished" podID="0eecc358-9581-489e-97ae-f600d35a7613" containerID="6f5e7b1c14376dcb0e62786390737b48cb04d2a05079d8160df0f9dedc56dfc8" exitCode=0 Jan 29 16:27:24 crc kubenswrapper[4714]: I0129 16:27:24.536261 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6" event={"ID":"0eecc358-9581-489e-97ae-f600d35a7613","Type":"ContainerDied","Data":"6f5e7b1c14376dcb0e62786390737b48cb04d2a05079d8160df0f9dedc56dfc8"} Jan 29 16:27:27 crc kubenswrapper[4714]: I0129 16:27:27.844279 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:27:27 crc kubenswrapper[4714]: I0129 16:27:27.845009 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:27:27 crc kubenswrapper[4714]: I0129 16:27:27.845152 4714 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:27:27 crc kubenswrapper[4714]: I0129 16:27:27.845793 4714 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"77045db0ac9dbee23fe648e58207222e15e50d5178fcc5cc7a606b4bbe2af7ec"} pod="openshift-machine-config-operator/machine-config-daemon-ppngk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 16:27:27 crc kubenswrapper[4714]: I0129 16:27:27.845842 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" containerID="cri-o://77045db0ac9dbee23fe648e58207222e15e50d5178fcc5cc7a606b4bbe2af7ec" gracePeriod=600 Jan 29 16:27:28 crc kubenswrapper[4714]: I0129 16:27:28.537973 4714 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 16:27:28 crc kubenswrapper[4714]: I0129 16:27:28.573826 4714 generic.go:334] "Generic (PLEG): container finished" podID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerID="77045db0ac9dbee23fe648e58207222e15e50d5178fcc5cc7a606b4bbe2af7ec" exitCode=0 Jan 29 16:27:28 crc kubenswrapper[4714]: I0129 16:27:28.573892 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" event={"ID":"c8c765f3-89eb-4077-8829-03e86eb0c90c","Type":"ContainerDied","Data":"77045db0ac9dbee23fe648e58207222e15e50d5178fcc5cc7a606b4bbe2af7ec"} Jan 29 16:27:28 crc kubenswrapper[4714]: I0129 16:27:28.574022 4714 scope.go:117] "RemoveContainer" containerID="434181b332ad91829c9ca3b07c475cac7d3c8b013492e90ce07fd88776d24efa" Jan 29 16:27:28 crc kubenswrapper[4714]: E0129 16:27:28.677064 4714 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 16:27:28 crc kubenswrapper[4714]: E0129 16:27:28.677336 4714 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tmm88,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-wdwq5_openshift-marketplace(8c12ad14-f878-42a1-a168-bad4026ec2dd): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 16:27:28 crc kubenswrapper[4714]: E0129 16:27:28.678693 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:27:29 crc kubenswrapper[4714]: I0129 16:27:29.581748 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" event={"ID":"c8c765f3-89eb-4077-8829-03e86eb0c90c","Type":"ContainerStarted","Data":"6d286411f160a5fdbd13efa6bbfae544ec01e44f19ea6b8ff05d4ab9953a5f4a"} Jan 29 16:27:29 crc kubenswrapper[4714]: I0129 16:27:29.583180 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-sync-6xxc6" event={"ID":"6e33ba3d-9561-441b-b835-fbdb6ce97d23","Type":"ContainerStarted","Data":"02c31eb5896b8dc80e78bcd830c0e2f150e33491e2a451310581ca2dc793d036"} Jan 29 16:27:29 crc kubenswrapper[4714]: I0129 16:27:29.585761 4714 generic.go:334] "Generic (PLEG): container finished" podID="0eecc358-9581-489e-97ae-f600d35a7613" containerID="8f755ca88cec23079c8fcc603a70054716d3829ccb1c6da9ff0f5feff88b5796" exitCode=0 Jan 29 16:27:29 crc kubenswrapper[4714]: I0129 16:27:29.585793 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6" event={"ID":"0eecc358-9581-489e-97ae-f600d35a7613","Type":"ContainerDied","Data":"8f755ca88cec23079c8fcc603a70054716d3829ccb1c6da9ff0f5feff88b5796"} Jan 29 16:27:29 crc kubenswrapper[4714]: I0129 16:27:29.611670 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone-db-sync-6xxc6" podStartSLOduration=1.8190485330000001 podStartE2EDuration="8.611652333s" podCreationTimestamp="2026-01-29 16:27:21 +0000 UTC" firstStartedPulling="2026-01-29 16:27:21.843499267 +0000 UTC m=+1048.364000387" lastFinishedPulling="2026-01-29 16:27:28.636103047 +0000 UTC m=+1055.156604187" observedRunningTime="2026-01-29 16:27:29.611333416 +0000 UTC m=+1056.131834546" watchObservedRunningTime="2026-01-29 16:27:29.611652333 +0000 UTC m=+1056.132153463" Jan 29 16:27:30 crc kubenswrapper[4714]: I0129 16:27:30.594257 4714 generic.go:334] "Generic (PLEG): container finished" podID="0eecc358-9581-489e-97ae-f600d35a7613" containerID="a675f4f90dd437578a32683cafb8b1908c7b80f63189aec46569a29c2add56c0" exitCode=0 Jan 29 16:27:30 crc kubenswrapper[4714]: I0129 16:27:30.594300 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6" event={"ID":"0eecc358-9581-489e-97ae-f600d35a7613","Type":"ContainerDied","Data":"a675f4f90dd437578a32683cafb8b1908c7b80f63189aec46569a29c2add56c0"} Jan 29 16:27:31 crc kubenswrapper[4714]: I0129 16:27:31.869042 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6" Jan 29 16:27:31 crc kubenswrapper[4714]: I0129 16:27:31.991845 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0eecc358-9581-489e-97ae-f600d35a7613-bundle\") pod \"0eecc358-9581-489e-97ae-f600d35a7613\" (UID: \"0eecc358-9581-489e-97ae-f600d35a7613\") " Jan 29 16:27:31 crc kubenswrapper[4714]: I0129 16:27:31.993056 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0eecc358-9581-489e-97ae-f600d35a7613-bundle" (OuterVolumeSpecName: "bundle") pod "0eecc358-9581-489e-97ae-f600d35a7613" (UID: "0eecc358-9581-489e-97ae-f600d35a7613"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:27:31 crc kubenswrapper[4714]: I0129 16:27:31.993237 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0eecc358-9581-489e-97ae-f600d35a7613-util\") pod \"0eecc358-9581-489e-97ae-f600d35a7613\" (UID: \"0eecc358-9581-489e-97ae-f600d35a7613\") " Jan 29 16:27:31 crc kubenswrapper[4714]: I0129 16:27:31.993416 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4wwh\" (UniqueName: \"kubernetes.io/projected/0eecc358-9581-489e-97ae-f600d35a7613-kube-api-access-k4wwh\") pod \"0eecc358-9581-489e-97ae-f600d35a7613\" (UID: \"0eecc358-9581-489e-97ae-f600d35a7613\") " Jan 29 16:27:31 crc kubenswrapper[4714]: I0129 16:27:31.994757 4714 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0eecc358-9581-489e-97ae-f600d35a7613-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:31 crc kubenswrapper[4714]: I0129 16:27:31.998998 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0eecc358-9581-489e-97ae-f600d35a7613-kube-api-access-k4wwh" (OuterVolumeSpecName: "kube-api-access-k4wwh") pod "0eecc358-9581-489e-97ae-f600d35a7613" (UID: "0eecc358-9581-489e-97ae-f600d35a7613"). InnerVolumeSpecName "kube-api-access-k4wwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:27:32 crc kubenswrapper[4714]: I0129 16:27:32.003493 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0eecc358-9581-489e-97ae-f600d35a7613-util" (OuterVolumeSpecName: "util") pod "0eecc358-9581-489e-97ae-f600d35a7613" (UID: "0eecc358-9581-489e-97ae-f600d35a7613"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:27:32 crc kubenswrapper[4714]: I0129 16:27:32.096231 4714 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0eecc358-9581-489e-97ae-f600d35a7613-util\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:32 crc kubenswrapper[4714]: I0129 16:27:32.096284 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4wwh\" (UniqueName: \"kubernetes.io/projected/0eecc358-9581-489e-97ae-f600d35a7613-kube-api-access-k4wwh\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:32 crc kubenswrapper[4714]: I0129 16:27:32.612601 4714 generic.go:334] "Generic (PLEG): container finished" podID="6e33ba3d-9561-441b-b835-fbdb6ce97d23" containerID="02c31eb5896b8dc80e78bcd830c0e2f150e33491e2a451310581ca2dc793d036" exitCode=0 Jan 29 16:27:32 crc kubenswrapper[4714]: I0129 16:27:32.612691 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-sync-6xxc6" event={"ID":"6e33ba3d-9561-441b-b835-fbdb6ce97d23","Type":"ContainerDied","Data":"02c31eb5896b8dc80e78bcd830c0e2f150e33491e2a451310581ca2dc793d036"} Jan 29 16:27:32 crc kubenswrapper[4714]: I0129 16:27:32.615790 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6" event={"ID":"0eecc358-9581-489e-97ae-f600d35a7613","Type":"ContainerDied","Data":"ccf8807833090dda1a00ca4aa3c347ca65828dd70468671dc4659950b5c10bcf"} Jan 29 16:27:32 crc kubenswrapper[4714]: I0129 16:27:32.615829 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccf8807833090dda1a00ca4aa3c347ca65828dd70468671dc4659950b5c10bcf" Jan 29 16:27:32 crc kubenswrapper[4714]: I0129 16:27:32.616189 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6" Jan 29 16:27:33 crc kubenswrapper[4714]: I0129 16:27:33.930059 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-sync-6xxc6" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.124480 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7s7m\" (UniqueName: \"kubernetes.io/projected/6e33ba3d-9561-441b-b835-fbdb6ce97d23-kube-api-access-f7s7m\") pod \"6e33ba3d-9561-441b-b835-fbdb6ce97d23\" (UID: \"6e33ba3d-9561-441b-b835-fbdb6ce97d23\") " Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.124544 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e33ba3d-9561-441b-b835-fbdb6ce97d23-config-data\") pod \"6e33ba3d-9561-441b-b835-fbdb6ce97d23\" (UID: \"6e33ba3d-9561-441b-b835-fbdb6ce97d23\") " Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.129919 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e33ba3d-9561-441b-b835-fbdb6ce97d23-kube-api-access-f7s7m" (OuterVolumeSpecName: "kube-api-access-f7s7m") pod "6e33ba3d-9561-441b-b835-fbdb6ce97d23" (UID: "6e33ba3d-9561-441b-b835-fbdb6ce97d23"). InnerVolumeSpecName "kube-api-access-f7s7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.154431 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e33ba3d-9561-441b-b835-fbdb6ce97d23-config-data" (OuterVolumeSpecName: "config-data") pod "6e33ba3d-9561-441b-b835-fbdb6ce97d23" (UID: "6e33ba3d-9561-441b-b835-fbdb6ce97d23"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.226411 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7s7m\" (UniqueName: \"kubernetes.io/projected/6e33ba3d-9561-441b-b835-fbdb6ce97d23-kube-api-access-f7s7m\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.226444 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e33ba3d-9561-441b-b835-fbdb6ce97d23-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.634538 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-sync-6xxc6" event={"ID":"6e33ba3d-9561-441b-b835-fbdb6ce97d23","Type":"ContainerDied","Data":"cb8e5c568c512500695d36838718c30771a365e70a1f7bf91812aeb8914126a0"} Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.634587 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb8e5c568c512500695d36838718c30771a365e70a1f7bf91812aeb8914126a0" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.634616 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-sync-6xxc6" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.805257 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-kj27d"] Jan 29 16:27:34 crc kubenswrapper[4714]: E0129 16:27:34.805502 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0eecc358-9581-489e-97ae-f600d35a7613" containerName="extract" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.805513 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="0eecc358-9581-489e-97ae-f600d35a7613" containerName="extract" Jan 29 16:27:34 crc kubenswrapper[4714]: E0129 16:27:34.805524 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0eecc358-9581-489e-97ae-f600d35a7613" containerName="pull" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.805529 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="0eecc358-9581-489e-97ae-f600d35a7613" containerName="pull" Jan 29 16:27:34 crc kubenswrapper[4714]: E0129 16:27:34.805546 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0eecc358-9581-489e-97ae-f600d35a7613" containerName="util" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.805552 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="0eecc358-9581-489e-97ae-f600d35a7613" containerName="util" Jan 29 16:27:34 crc kubenswrapper[4714]: E0129 16:27:34.805562 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e33ba3d-9561-441b-b835-fbdb6ce97d23" containerName="keystone-db-sync" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.805568 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e33ba3d-9561-441b-b835-fbdb6ce97d23" containerName="keystone-db-sync" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.805672 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="0eecc358-9581-489e-97ae-f600d35a7613" containerName="extract" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.805680 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e33ba3d-9561-441b-b835-fbdb6ce97d23" containerName="keystone-db-sync" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.806114 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.809094 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-keystone-dockercfg-6m228" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.809561 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-scripts" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.810253 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"osp-secret" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.811759 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.812037 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-config-data" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.815314 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-kj27d"] Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.835517 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-credential-keys\") pod \"keystone-bootstrap-kj27d\" (UID: \"f3e03982-d953-488f-a01a-5024f64ad7da\") " pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.835566 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-config-data\") pod \"keystone-bootstrap-kj27d\" (UID: \"f3e03982-d953-488f-a01a-5024f64ad7da\") " pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.835586 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-fernet-keys\") pod \"keystone-bootstrap-kj27d\" (UID: \"f3e03982-d953-488f-a01a-5024f64ad7da\") " pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.835614 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7nmp\" (UniqueName: \"kubernetes.io/projected/f3e03982-d953-488f-a01a-5024f64ad7da-kube-api-access-x7nmp\") pod \"keystone-bootstrap-kj27d\" (UID: \"f3e03982-d953-488f-a01a-5024f64ad7da\") " pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.835651 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-scripts\") pod \"keystone-bootstrap-kj27d\" (UID: \"f3e03982-d953-488f-a01a-5024f64ad7da\") " pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.936684 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-credential-keys\") pod \"keystone-bootstrap-kj27d\" (UID: \"f3e03982-d953-488f-a01a-5024f64ad7da\") " pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.936757 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-config-data\") pod \"keystone-bootstrap-kj27d\" (UID: \"f3e03982-d953-488f-a01a-5024f64ad7da\") " pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.936780 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-fernet-keys\") pod \"keystone-bootstrap-kj27d\" (UID: \"f3e03982-d953-488f-a01a-5024f64ad7da\") " pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.936824 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7nmp\" (UniqueName: \"kubernetes.io/projected/f3e03982-d953-488f-a01a-5024f64ad7da-kube-api-access-x7nmp\") pod \"keystone-bootstrap-kj27d\" (UID: \"f3e03982-d953-488f-a01a-5024f64ad7da\") " pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.937623 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-scripts\") pod \"keystone-bootstrap-kj27d\" (UID: \"f3e03982-d953-488f-a01a-5024f64ad7da\") " pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.942671 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-credential-keys\") pod \"keystone-bootstrap-kj27d\" (UID: \"f3e03982-d953-488f-a01a-5024f64ad7da\") " pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.943583 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-fernet-keys\") pod \"keystone-bootstrap-kj27d\" (UID: \"f3e03982-d953-488f-a01a-5024f64ad7da\") " pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.944859 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-config-data\") pod \"keystone-bootstrap-kj27d\" (UID: \"f3e03982-d953-488f-a01a-5024f64ad7da\") " pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.946793 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-scripts\") pod \"keystone-bootstrap-kj27d\" (UID: \"f3e03982-d953-488f-a01a-5024f64ad7da\") " pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" Jan 29 16:27:34 crc kubenswrapper[4714]: I0129 16:27:34.963556 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7nmp\" (UniqueName: \"kubernetes.io/projected/f3e03982-d953-488f-a01a-5024f64ad7da-kube-api-access-x7nmp\") pod \"keystone-bootstrap-kj27d\" (UID: \"f3e03982-d953-488f-a01a-5024f64ad7da\") " pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" Jan 29 16:27:35 crc kubenswrapper[4714]: I0129 16:27:35.131731 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" Jan 29 16:27:35 crc kubenswrapper[4714]: I0129 16:27:35.597879 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-kj27d"] Jan 29 16:27:35 crc kubenswrapper[4714]: I0129 16:27:35.663405 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" event={"ID":"f3e03982-d953-488f-a01a-5024f64ad7da","Type":"ContainerStarted","Data":"a624e5601997570e21553a0e88b80fa0db701686457a6ca234e4cc2c3067c658"} Jan 29 16:27:36 crc kubenswrapper[4714]: I0129 16:27:36.671170 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" event={"ID":"f3e03982-d953-488f-a01a-5024f64ad7da","Type":"ContainerStarted","Data":"6cf56f6dac5db6cefc7926b5a24bd8c2963224c5d6e15dd78662ec20f0cf0141"} Jan 29 16:27:36 crc kubenswrapper[4714]: I0129 16:27:36.688491 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" podStartSLOduration=2.688476602 podStartE2EDuration="2.688476602s" podCreationTimestamp="2026-01-29 16:27:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:27:36.686893358 +0000 UTC m=+1063.207394478" watchObservedRunningTime="2026-01-29 16:27:36.688476602 +0000 UTC m=+1063.208977722" Jan 29 16:27:38 crc kubenswrapper[4714]: I0129 16:27:38.686219 4714 generic.go:334] "Generic (PLEG): container finished" podID="f3e03982-d953-488f-a01a-5024f64ad7da" containerID="6cf56f6dac5db6cefc7926b5a24bd8c2963224c5d6e15dd78662ec20f0cf0141" exitCode=0 Jan 29 16:27:38 crc kubenswrapper[4714]: I0129 16:27:38.686262 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" event={"ID":"f3e03982-d953-488f-a01a-5024f64ad7da","Type":"ContainerDied","Data":"6cf56f6dac5db6cefc7926b5a24bd8c2963224c5d6e15dd78662ec20f0cf0141"} Jan 29 16:27:39 crc kubenswrapper[4714]: I0129 16:27:39.970498 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.117458 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7nmp\" (UniqueName: \"kubernetes.io/projected/f3e03982-d953-488f-a01a-5024f64ad7da-kube-api-access-x7nmp\") pod \"f3e03982-d953-488f-a01a-5024f64ad7da\" (UID: \"f3e03982-d953-488f-a01a-5024f64ad7da\") " Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.117551 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-config-data\") pod \"f3e03982-d953-488f-a01a-5024f64ad7da\" (UID: \"f3e03982-d953-488f-a01a-5024f64ad7da\") " Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.121795 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-scripts\") pod \"f3e03982-d953-488f-a01a-5024f64ad7da\" (UID: \"f3e03982-d953-488f-a01a-5024f64ad7da\") " Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.122260 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-credential-keys\") pod \"f3e03982-d953-488f-a01a-5024f64ad7da\" (UID: \"f3e03982-d953-488f-a01a-5024f64ad7da\") " Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.122452 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-fernet-keys\") pod \"f3e03982-d953-488f-a01a-5024f64ad7da\" (UID: \"f3e03982-d953-488f-a01a-5024f64ad7da\") " Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.125179 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-scripts" (OuterVolumeSpecName: "scripts") pod "f3e03982-d953-488f-a01a-5024f64ad7da" (UID: "f3e03982-d953-488f-a01a-5024f64ad7da"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.125300 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3e03982-d953-488f-a01a-5024f64ad7da-kube-api-access-x7nmp" (OuterVolumeSpecName: "kube-api-access-x7nmp") pod "f3e03982-d953-488f-a01a-5024f64ad7da" (UID: "f3e03982-d953-488f-a01a-5024f64ad7da"). InnerVolumeSpecName "kube-api-access-x7nmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.142173 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f3e03982-d953-488f-a01a-5024f64ad7da" (UID: "f3e03982-d953-488f-a01a-5024f64ad7da"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.143406 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "f3e03982-d953-488f-a01a-5024f64ad7da" (UID: "f3e03982-d953-488f-a01a-5024f64ad7da"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.153681 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-config-data" (OuterVolumeSpecName: "config-data") pod "f3e03982-d953-488f-a01a-5024f64ad7da" (UID: "f3e03982-d953-488f-a01a-5024f64ad7da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.224725 4714 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.224764 4714 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.224776 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7nmp\" (UniqueName: \"kubernetes.io/projected/f3e03982-d953-488f-a01a-5024f64ad7da-kube-api-access-x7nmp\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.224787 4714 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.224798 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3e03982-d953-488f-a01a-5024f64ad7da-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.698251 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" event={"ID":"f3e03982-d953-488f-a01a-5024f64ad7da","Type":"ContainerDied","Data":"a624e5601997570e21553a0e88b80fa0db701686457a6ca234e4cc2c3067c658"} Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.698538 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a624e5601997570e21553a0e88b80fa0db701686457a6ca234e4cc2c3067c658" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.698311 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-bootstrap-kj27d" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.863686 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-db9b49999-6gd95"] Jan 29 16:27:40 crc kubenswrapper[4714]: E0129 16:27:40.863948 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3e03982-d953-488f-a01a-5024f64ad7da" containerName="keystone-bootstrap" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.863964 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3e03982-d953-488f-a01a-5024f64ad7da" containerName="keystone-bootstrap" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.864085 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3e03982-d953-488f-a01a-5024f64ad7da" containerName="keystone-bootstrap" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.864515 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.866386 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-scripts" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.866491 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-keystone-dockercfg-6m228" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.866496 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.867197 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-config-data" Jan 29 16:27:40 crc kubenswrapper[4714]: I0129 16:27:40.884596 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db9b49999-6gd95"] Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.036589 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-scripts\") pod \"keystone-db9b49999-6gd95\" (UID: \"fc06a535-6f60-438e-b52d-5dc90fae8c67\") " pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.036692 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-fernet-keys\") pod \"keystone-db9b49999-6gd95\" (UID: \"fc06a535-6f60-438e-b52d-5dc90fae8c67\") " pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.036729 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-credential-keys\") pod \"keystone-db9b49999-6gd95\" (UID: \"fc06a535-6f60-438e-b52d-5dc90fae8c67\") " pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.036761 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2vn9\" (UniqueName: \"kubernetes.io/projected/fc06a535-6f60-438e-b52d-5dc90fae8c67-kube-api-access-c2vn9\") pod \"keystone-db9b49999-6gd95\" (UID: \"fc06a535-6f60-438e-b52d-5dc90fae8c67\") " pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.036823 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-config-data\") pod \"keystone-db9b49999-6gd95\" (UID: \"fc06a535-6f60-438e-b52d-5dc90fae8c67\") " pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.138536 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-config-data\") pod \"keystone-db9b49999-6gd95\" (UID: \"fc06a535-6f60-438e-b52d-5dc90fae8c67\") " pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.138606 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-scripts\") pod \"keystone-db9b49999-6gd95\" (UID: \"fc06a535-6f60-438e-b52d-5dc90fae8c67\") " pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.138674 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-fernet-keys\") pod \"keystone-db9b49999-6gd95\" (UID: \"fc06a535-6f60-438e-b52d-5dc90fae8c67\") " pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.138714 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-credential-keys\") pod \"keystone-db9b49999-6gd95\" (UID: \"fc06a535-6f60-438e-b52d-5dc90fae8c67\") " pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.138744 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2vn9\" (UniqueName: \"kubernetes.io/projected/fc06a535-6f60-438e-b52d-5dc90fae8c67-kube-api-access-c2vn9\") pod \"keystone-db9b49999-6gd95\" (UID: \"fc06a535-6f60-438e-b52d-5dc90fae8c67\") " pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.144107 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-fernet-keys\") pod \"keystone-db9b49999-6gd95\" (UID: \"fc06a535-6f60-438e-b52d-5dc90fae8c67\") " pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.144125 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-config-data\") pod \"keystone-db9b49999-6gd95\" (UID: \"fc06a535-6f60-438e-b52d-5dc90fae8c67\") " pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.144398 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-scripts\") pod \"keystone-db9b49999-6gd95\" (UID: \"fc06a535-6f60-438e-b52d-5dc90fae8c67\") " pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.144832 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-credential-keys\") pod \"keystone-db9b49999-6gd95\" (UID: \"fc06a535-6f60-438e-b52d-5dc90fae8c67\") " pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.161181 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2vn9\" (UniqueName: \"kubernetes.io/projected/fc06a535-6f60-438e-b52d-5dc90fae8c67-kube-api-access-c2vn9\") pod \"keystone-db9b49999-6gd95\" (UID: \"fc06a535-6f60-438e-b52d-5dc90fae8c67\") " pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.182176 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:27:41 crc kubenswrapper[4714]: E0129 16:27:41.185372 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.472098 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs"] Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.472909 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.475172 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-service-cert" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.475200 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-72wx2" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.495820 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs"] Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.643966 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9aa31790-7a3c-4a66-aace-c087c0221c6b-webhook-cert\") pod \"cinder-operator-controller-manager-5fc6d4b6f5-9mdcs\" (UID: \"9aa31790-7a3c-4a66-aace-c087c0221c6b\") " pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.644015 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4dvn\" (UniqueName: \"kubernetes.io/projected/9aa31790-7a3c-4a66-aace-c087c0221c6b-kube-api-access-t4dvn\") pod \"cinder-operator-controller-manager-5fc6d4b6f5-9mdcs\" (UID: \"9aa31790-7a3c-4a66-aace-c087c0221c6b\") " pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.644062 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9aa31790-7a3c-4a66-aace-c087c0221c6b-apiservice-cert\") pod \"cinder-operator-controller-manager-5fc6d4b6f5-9mdcs\" (UID: \"9aa31790-7a3c-4a66-aace-c087c0221c6b\") " pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.647497 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db9b49999-6gd95"] Jan 29 16:27:41 crc kubenswrapper[4714]: W0129 16:27:41.648476 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc06a535_6f60_438e_b52d_5dc90fae8c67.slice/crio-a35aebf0427b3f34153f6b20222d8016725e500d95a3072063dc0d02bd8d902e WatchSource:0}: Error finding container a35aebf0427b3f34153f6b20222d8016725e500d95a3072063dc0d02bd8d902e: Status 404 returned error can't find the container with id a35aebf0427b3f34153f6b20222d8016725e500d95a3072063dc0d02bd8d902e Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.705404 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" event={"ID":"fc06a535-6f60-438e-b52d-5dc90fae8c67","Type":"ContainerStarted","Data":"a35aebf0427b3f34153f6b20222d8016725e500d95a3072063dc0d02bd8d902e"} Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.744984 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9aa31790-7a3c-4a66-aace-c087c0221c6b-webhook-cert\") pod \"cinder-operator-controller-manager-5fc6d4b6f5-9mdcs\" (UID: \"9aa31790-7a3c-4a66-aace-c087c0221c6b\") " pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.745034 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4dvn\" (UniqueName: \"kubernetes.io/projected/9aa31790-7a3c-4a66-aace-c087c0221c6b-kube-api-access-t4dvn\") pod \"cinder-operator-controller-manager-5fc6d4b6f5-9mdcs\" (UID: \"9aa31790-7a3c-4a66-aace-c087c0221c6b\") " pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.745073 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9aa31790-7a3c-4a66-aace-c087c0221c6b-apiservice-cert\") pod \"cinder-operator-controller-manager-5fc6d4b6f5-9mdcs\" (UID: \"9aa31790-7a3c-4a66-aace-c087c0221c6b\") " pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.749804 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9aa31790-7a3c-4a66-aace-c087c0221c6b-webhook-cert\") pod \"cinder-operator-controller-manager-5fc6d4b6f5-9mdcs\" (UID: \"9aa31790-7a3c-4a66-aace-c087c0221c6b\") " pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.753586 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9aa31790-7a3c-4a66-aace-c087c0221c6b-apiservice-cert\") pod \"cinder-operator-controller-manager-5fc6d4b6f5-9mdcs\" (UID: \"9aa31790-7a3c-4a66-aace-c087c0221c6b\") " pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.764680 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4dvn\" (UniqueName: \"kubernetes.io/projected/9aa31790-7a3c-4a66-aace-c087c0221c6b-kube-api-access-t4dvn\") pod \"cinder-operator-controller-manager-5fc6d4b6f5-9mdcs\" (UID: \"9aa31790-7a3c-4a66-aace-c087c0221c6b\") " pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" Jan 29 16:27:41 crc kubenswrapper[4714]: I0129 16:27:41.795655 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" Jan 29 16:27:42 crc kubenswrapper[4714]: I0129 16:27:42.238622 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs"] Jan 29 16:27:42 crc kubenswrapper[4714]: W0129 16:27:42.243561 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9aa31790_7a3c_4a66_aace_c087c0221c6b.slice/crio-c82d440d4556664aec8a776ead06b3d925538dac3151f1c3b85d4cf089d48d43 WatchSource:0}: Error finding container c82d440d4556664aec8a776ead06b3d925538dac3151f1c3b85d4cf089d48d43: Status 404 returned error can't find the container with id c82d440d4556664aec8a776ead06b3d925538dac3151f1c3b85d4cf089d48d43 Jan 29 16:27:42 crc kubenswrapper[4714]: I0129 16:27:42.714635 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" event={"ID":"9aa31790-7a3c-4a66-aace-c087c0221c6b","Type":"ContainerStarted","Data":"c82d440d4556664aec8a776ead06b3d925538dac3151f1c3b85d4cf089d48d43"} Jan 29 16:27:42 crc kubenswrapper[4714]: I0129 16:27:42.716488 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" event={"ID":"fc06a535-6f60-438e-b52d-5dc90fae8c67","Type":"ContainerStarted","Data":"4adeb441736bb7d27c549844b9147979d27f5dbed7e42a9b990619d275a00000"} Jan 29 16:27:42 crc kubenswrapper[4714]: I0129 16:27:42.716942 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:27:42 crc kubenswrapper[4714]: I0129 16:27:42.742143 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" podStartSLOduration=2.742116393 podStartE2EDuration="2.742116393s" podCreationTimestamp="2026-01-29 16:27:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:27:42.733188242 +0000 UTC m=+1069.253689362" watchObservedRunningTime="2026-01-29 16:27:42.742116393 +0000 UTC m=+1069.262617513" Jan 29 16:27:45 crc kubenswrapper[4714]: I0129 16:27:45.748701 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" event={"ID":"9aa31790-7a3c-4a66-aace-c087c0221c6b","Type":"ContainerStarted","Data":"36513d8368599e830d1d92fc041ab07bd030a4ddf286f084b21d04377e2a5dbd"} Jan 29 16:27:45 crc kubenswrapper[4714]: I0129 16:27:45.749604 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" Jan 29 16:27:45 crc kubenswrapper[4714]: I0129 16:27:45.765324 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" podStartSLOduration=2.200486921 podStartE2EDuration="4.765307912s" podCreationTimestamp="2026-01-29 16:27:41 +0000 UTC" firstStartedPulling="2026-01-29 16:27:42.245788463 +0000 UTC m=+1068.766289583" lastFinishedPulling="2026-01-29 16:27:44.810609454 +0000 UTC m=+1071.331110574" observedRunningTime="2026-01-29 16:27:45.761414859 +0000 UTC m=+1072.281915979" watchObservedRunningTime="2026-01-29 16:27:45.765307912 +0000 UTC m=+1072.285809022" Jan 29 16:27:51 crc kubenswrapper[4714]: I0129 16:27:51.800173 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" Jan 29 16:27:52 crc kubenswrapper[4714]: E0129 16:27:52.186568 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.163894 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-db-create-64s46"] Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.164934 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-64s46" Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.170025 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-9de3-account-create-update-79889"] Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.170828 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-9de3-account-create-update-79889" Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.172732 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-db-secret" Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.177592 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-64s46"] Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.184461 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-9de3-account-create-update-79889"] Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.233436 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rznp\" (UniqueName: \"kubernetes.io/projected/e670557b-650e-478c-9f87-eaba6641f02f-kube-api-access-4rznp\") pod \"cinder-9de3-account-create-update-79889\" (UID: \"e670557b-650e-478c-9f87-eaba6641f02f\") " pod="cinder-kuttl-tests/cinder-9de3-account-create-update-79889" Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.233498 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c747693b-f2e9-4073-9432-115643a6b6d1-operator-scripts\") pod \"cinder-db-create-64s46\" (UID: \"c747693b-f2e9-4073-9432-115643a6b6d1\") " pod="cinder-kuttl-tests/cinder-db-create-64s46" Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.233596 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e670557b-650e-478c-9f87-eaba6641f02f-operator-scripts\") pod \"cinder-9de3-account-create-update-79889\" (UID: \"e670557b-650e-478c-9f87-eaba6641f02f\") " pod="cinder-kuttl-tests/cinder-9de3-account-create-update-79889" Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.233627 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxhv6\" (UniqueName: \"kubernetes.io/projected/c747693b-f2e9-4073-9432-115643a6b6d1-kube-api-access-fxhv6\") pod \"cinder-db-create-64s46\" (UID: \"c747693b-f2e9-4073-9432-115643a6b6d1\") " pod="cinder-kuttl-tests/cinder-db-create-64s46" Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.334793 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rznp\" (UniqueName: \"kubernetes.io/projected/e670557b-650e-478c-9f87-eaba6641f02f-kube-api-access-4rznp\") pod \"cinder-9de3-account-create-update-79889\" (UID: \"e670557b-650e-478c-9f87-eaba6641f02f\") " pod="cinder-kuttl-tests/cinder-9de3-account-create-update-79889" Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.335248 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c747693b-f2e9-4073-9432-115643a6b6d1-operator-scripts\") pod \"cinder-db-create-64s46\" (UID: \"c747693b-f2e9-4073-9432-115643a6b6d1\") " pod="cinder-kuttl-tests/cinder-db-create-64s46" Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.335911 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c747693b-f2e9-4073-9432-115643a6b6d1-operator-scripts\") pod \"cinder-db-create-64s46\" (UID: \"c747693b-f2e9-4073-9432-115643a6b6d1\") " pod="cinder-kuttl-tests/cinder-db-create-64s46" Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.336053 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e670557b-650e-478c-9f87-eaba6641f02f-operator-scripts\") pod \"cinder-9de3-account-create-update-79889\" (UID: \"e670557b-650e-478c-9f87-eaba6641f02f\") " pod="cinder-kuttl-tests/cinder-9de3-account-create-update-79889" Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.336684 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e670557b-650e-478c-9f87-eaba6641f02f-operator-scripts\") pod \"cinder-9de3-account-create-update-79889\" (UID: \"e670557b-650e-478c-9f87-eaba6641f02f\") " pod="cinder-kuttl-tests/cinder-9de3-account-create-update-79889" Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.336751 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxhv6\" (UniqueName: \"kubernetes.io/projected/c747693b-f2e9-4073-9432-115643a6b6d1-kube-api-access-fxhv6\") pod \"cinder-db-create-64s46\" (UID: \"c747693b-f2e9-4073-9432-115643a6b6d1\") " pod="cinder-kuttl-tests/cinder-db-create-64s46" Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.359737 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rznp\" (UniqueName: \"kubernetes.io/projected/e670557b-650e-478c-9f87-eaba6641f02f-kube-api-access-4rznp\") pod \"cinder-9de3-account-create-update-79889\" (UID: \"e670557b-650e-478c-9f87-eaba6641f02f\") " pod="cinder-kuttl-tests/cinder-9de3-account-create-update-79889" Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.376498 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxhv6\" (UniqueName: \"kubernetes.io/projected/c747693b-f2e9-4073-9432-115643a6b6d1-kube-api-access-fxhv6\") pod \"cinder-db-create-64s46\" (UID: \"c747693b-f2e9-4073-9432-115643a6b6d1\") " pod="cinder-kuttl-tests/cinder-db-create-64s46" Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.481701 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-64s46" Jan 29 16:27:55 crc kubenswrapper[4714]: I0129 16:27:55.500435 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-9de3-account-create-update-79889" Jan 29 16:27:56 crc kubenswrapper[4714]: W0129 16:27:56.008208 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc747693b_f2e9_4073_9432_115643a6b6d1.slice/crio-ef09ca095d654272ce164d737756bdd07a31ec4b52f6796143d33c3b42a8fa27 WatchSource:0}: Error finding container ef09ca095d654272ce164d737756bdd07a31ec4b52f6796143d33c3b42a8fa27: Status 404 returned error can't find the container with id ef09ca095d654272ce164d737756bdd07a31ec4b52f6796143d33c3b42a8fa27 Jan 29 16:27:56 crc kubenswrapper[4714]: I0129 16:27:56.010465 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-64s46"] Jan 29 16:27:56 crc kubenswrapper[4714]: I0129 16:27:56.049747 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-9de3-account-create-update-79889"] Jan 29 16:27:56 crc kubenswrapper[4714]: W0129 16:27:56.050716 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode670557b_650e_478c_9f87_eaba6641f02f.slice/crio-326a8c0db24835df9017b544cd8815b14dee4cd8df1be6f526c72cc4768020f6 WatchSource:0}: Error finding container 326a8c0db24835df9017b544cd8815b14dee4cd8df1be6f526c72cc4768020f6: Status 404 returned error can't find the container with id 326a8c0db24835df9017b544cd8815b14dee4cd8df1be6f526c72cc4768020f6 Jan 29 16:27:56 crc kubenswrapper[4714]: I0129 16:27:56.840621 4714 generic.go:334] "Generic (PLEG): container finished" podID="c747693b-f2e9-4073-9432-115643a6b6d1" containerID="cd8982aadd49edb0050578e4754c053be8b4e593ef390dafa6884ec2cec1fd5d" exitCode=0 Jan 29 16:27:56 crc kubenswrapper[4714]: I0129 16:27:56.840674 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-64s46" event={"ID":"c747693b-f2e9-4073-9432-115643a6b6d1","Type":"ContainerDied","Data":"cd8982aadd49edb0050578e4754c053be8b4e593ef390dafa6884ec2cec1fd5d"} Jan 29 16:27:56 crc kubenswrapper[4714]: I0129 16:27:56.841003 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-64s46" event={"ID":"c747693b-f2e9-4073-9432-115643a6b6d1","Type":"ContainerStarted","Data":"ef09ca095d654272ce164d737756bdd07a31ec4b52f6796143d33c3b42a8fa27"} Jan 29 16:27:56 crc kubenswrapper[4714]: I0129 16:27:56.842472 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-9de3-account-create-update-79889" event={"ID":"e670557b-650e-478c-9f87-eaba6641f02f","Type":"ContainerStarted","Data":"bc9178e686ab88b7e47825dd5faad2c6f1b972c479a40bdd9847878e376e9b8c"} Jan 29 16:27:56 crc kubenswrapper[4714]: I0129 16:27:56.842545 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-9de3-account-create-update-79889" event={"ID":"e670557b-650e-478c-9f87-eaba6641f02f","Type":"ContainerStarted","Data":"326a8c0db24835df9017b544cd8815b14dee4cd8df1be6f526c72cc4768020f6"} Jan 29 16:27:56 crc kubenswrapper[4714]: I0129 16:27:56.894591 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-9de3-account-create-update-79889" podStartSLOduration=1.894569857 podStartE2EDuration="1.894569857s" podCreationTimestamp="2026-01-29 16:27:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:27:56.889322564 +0000 UTC m=+1083.409823684" watchObservedRunningTime="2026-01-29 16:27:56.894569857 +0000 UTC m=+1083.415070977" Jan 29 16:27:57 crc kubenswrapper[4714]: I0129 16:27:57.850605 4714 generic.go:334] "Generic (PLEG): container finished" podID="e670557b-650e-478c-9f87-eaba6641f02f" containerID="bc9178e686ab88b7e47825dd5faad2c6f1b972c479a40bdd9847878e376e9b8c" exitCode=0 Jan 29 16:27:57 crc kubenswrapper[4714]: I0129 16:27:57.850698 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-9de3-account-create-update-79889" event={"ID":"e670557b-650e-478c-9f87-eaba6641f02f","Type":"ContainerDied","Data":"bc9178e686ab88b7e47825dd5faad2c6f1b972c479a40bdd9847878e376e9b8c"} Jan 29 16:27:58 crc kubenswrapper[4714]: I0129 16:27:58.214235 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-64s46" Jan 29 16:27:58 crc kubenswrapper[4714]: I0129 16:27:58.376486 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxhv6\" (UniqueName: \"kubernetes.io/projected/c747693b-f2e9-4073-9432-115643a6b6d1-kube-api-access-fxhv6\") pod \"c747693b-f2e9-4073-9432-115643a6b6d1\" (UID: \"c747693b-f2e9-4073-9432-115643a6b6d1\") " Jan 29 16:27:58 crc kubenswrapper[4714]: I0129 16:27:58.376533 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c747693b-f2e9-4073-9432-115643a6b6d1-operator-scripts\") pod \"c747693b-f2e9-4073-9432-115643a6b6d1\" (UID: \"c747693b-f2e9-4073-9432-115643a6b6d1\") " Jan 29 16:27:58 crc kubenswrapper[4714]: I0129 16:27:58.377622 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c747693b-f2e9-4073-9432-115643a6b6d1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c747693b-f2e9-4073-9432-115643a6b6d1" (UID: "c747693b-f2e9-4073-9432-115643a6b6d1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:27:58 crc kubenswrapper[4714]: I0129 16:27:58.385952 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c747693b-f2e9-4073-9432-115643a6b6d1-kube-api-access-fxhv6" (OuterVolumeSpecName: "kube-api-access-fxhv6") pod "c747693b-f2e9-4073-9432-115643a6b6d1" (UID: "c747693b-f2e9-4073-9432-115643a6b6d1"). InnerVolumeSpecName "kube-api-access-fxhv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:27:58 crc kubenswrapper[4714]: I0129 16:27:58.478576 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxhv6\" (UniqueName: \"kubernetes.io/projected/c747693b-f2e9-4073-9432-115643a6b6d1-kube-api-access-fxhv6\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:58 crc kubenswrapper[4714]: I0129 16:27:58.478614 4714 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c747693b-f2e9-4073-9432-115643a6b6d1-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:58 crc kubenswrapper[4714]: I0129 16:27:58.858548 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-64s46" Jan 29 16:27:58 crc kubenswrapper[4714]: I0129 16:27:58.858548 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-64s46" event={"ID":"c747693b-f2e9-4073-9432-115643a6b6d1","Type":"ContainerDied","Data":"ef09ca095d654272ce164d737756bdd07a31ec4b52f6796143d33c3b42a8fa27"} Jan 29 16:27:58 crc kubenswrapper[4714]: I0129 16:27:58.858972 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef09ca095d654272ce164d737756bdd07a31ec4b52f6796143d33c3b42a8fa27" Jan 29 16:27:59 crc kubenswrapper[4714]: I0129 16:27:59.134146 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-9de3-account-create-update-79889" Jan 29 16:27:59 crc kubenswrapper[4714]: I0129 16:27:59.287396 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e670557b-650e-478c-9f87-eaba6641f02f-operator-scripts\") pod \"e670557b-650e-478c-9f87-eaba6641f02f\" (UID: \"e670557b-650e-478c-9f87-eaba6641f02f\") " Jan 29 16:27:59 crc kubenswrapper[4714]: I0129 16:27:59.287504 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rznp\" (UniqueName: \"kubernetes.io/projected/e670557b-650e-478c-9f87-eaba6641f02f-kube-api-access-4rznp\") pod \"e670557b-650e-478c-9f87-eaba6641f02f\" (UID: \"e670557b-650e-478c-9f87-eaba6641f02f\") " Jan 29 16:27:59 crc kubenswrapper[4714]: I0129 16:27:59.288470 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e670557b-650e-478c-9f87-eaba6641f02f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e670557b-650e-478c-9f87-eaba6641f02f" (UID: "e670557b-650e-478c-9f87-eaba6641f02f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:27:59 crc kubenswrapper[4714]: I0129 16:27:59.298088 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e670557b-650e-478c-9f87-eaba6641f02f-kube-api-access-4rznp" (OuterVolumeSpecName: "kube-api-access-4rznp") pod "e670557b-650e-478c-9f87-eaba6641f02f" (UID: "e670557b-650e-478c-9f87-eaba6641f02f"). InnerVolumeSpecName "kube-api-access-4rznp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:27:59 crc kubenswrapper[4714]: I0129 16:27:59.389619 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rznp\" (UniqueName: \"kubernetes.io/projected/e670557b-650e-478c-9f87-eaba6641f02f-kube-api-access-4rznp\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:59 crc kubenswrapper[4714]: I0129 16:27:59.389896 4714 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e670557b-650e-478c-9f87-eaba6641f02f-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:27:59 crc kubenswrapper[4714]: I0129 16:27:59.866798 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-9de3-account-create-update-79889" event={"ID":"e670557b-650e-478c-9f87-eaba6641f02f","Type":"ContainerDied","Data":"326a8c0db24835df9017b544cd8815b14dee4cd8df1be6f526c72cc4768020f6"} Jan 29 16:27:59 crc kubenswrapper[4714]: I0129 16:27:59.866839 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="326a8c0db24835df9017b544cd8815b14dee4cd8df1be6f526c72cc4768020f6" Jan 29 16:27:59 crc kubenswrapper[4714]: I0129 16:27:59.866866 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-9de3-account-create-update-79889" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.508137 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-9pvrg"] Jan 29 16:28:00 crc kubenswrapper[4714]: E0129 16:28:00.508432 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c747693b-f2e9-4073-9432-115643a6b6d1" containerName="mariadb-database-create" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.508454 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="c747693b-f2e9-4073-9432-115643a6b6d1" containerName="mariadb-database-create" Jan 29 16:28:00 crc kubenswrapper[4714]: E0129 16:28:00.508482 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e670557b-650e-478c-9f87-eaba6641f02f" containerName="mariadb-account-create-update" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.508492 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="e670557b-650e-478c-9f87-eaba6641f02f" containerName="mariadb-account-create-update" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.508631 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="e670557b-650e-478c-9f87-eaba6641f02f" containerName="mariadb-account-create-update" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.508645 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="c747693b-f2e9-4073-9432-115643a6b6d1" containerName="mariadb-database-create" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.509127 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.514877 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-cinder-dockercfg-xw2vq" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.515154 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scripts" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.515452 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-config-data" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.531365 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-9pvrg"] Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.606277 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcnsl\" (UniqueName: \"kubernetes.io/projected/f09789c2-52ed-4321-95f8-02c3b3f271e3-kube-api-access-hcnsl\") pod \"cinder-db-sync-9pvrg\" (UID: \"f09789c2-52ed-4321-95f8-02c3b3f271e3\") " pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.606330 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f09789c2-52ed-4321-95f8-02c3b3f271e3-etc-machine-id\") pod \"cinder-db-sync-9pvrg\" (UID: \"f09789c2-52ed-4321-95f8-02c3b3f271e3\") " pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.606436 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f09789c2-52ed-4321-95f8-02c3b3f271e3-config-data\") pod \"cinder-db-sync-9pvrg\" (UID: \"f09789c2-52ed-4321-95f8-02c3b3f271e3\") " pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.606484 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f09789c2-52ed-4321-95f8-02c3b3f271e3-scripts\") pod \"cinder-db-sync-9pvrg\" (UID: \"f09789c2-52ed-4321-95f8-02c3b3f271e3\") " pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.606661 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f09789c2-52ed-4321-95f8-02c3b3f271e3-db-sync-config-data\") pod \"cinder-db-sync-9pvrg\" (UID: \"f09789c2-52ed-4321-95f8-02c3b3f271e3\") " pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.707892 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcnsl\" (UniqueName: \"kubernetes.io/projected/f09789c2-52ed-4321-95f8-02c3b3f271e3-kube-api-access-hcnsl\") pod \"cinder-db-sync-9pvrg\" (UID: \"f09789c2-52ed-4321-95f8-02c3b3f271e3\") " pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.708006 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f09789c2-52ed-4321-95f8-02c3b3f271e3-etc-machine-id\") pod \"cinder-db-sync-9pvrg\" (UID: \"f09789c2-52ed-4321-95f8-02c3b3f271e3\") " pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.708058 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f09789c2-52ed-4321-95f8-02c3b3f271e3-config-data\") pod \"cinder-db-sync-9pvrg\" (UID: \"f09789c2-52ed-4321-95f8-02c3b3f271e3\") " pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.708085 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f09789c2-52ed-4321-95f8-02c3b3f271e3-scripts\") pod \"cinder-db-sync-9pvrg\" (UID: \"f09789c2-52ed-4321-95f8-02c3b3f271e3\") " pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.708151 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f09789c2-52ed-4321-95f8-02c3b3f271e3-db-sync-config-data\") pod \"cinder-db-sync-9pvrg\" (UID: \"f09789c2-52ed-4321-95f8-02c3b3f271e3\") " pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.708084 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f09789c2-52ed-4321-95f8-02c3b3f271e3-etc-machine-id\") pod \"cinder-db-sync-9pvrg\" (UID: \"f09789c2-52ed-4321-95f8-02c3b3f271e3\") " pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.725720 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f09789c2-52ed-4321-95f8-02c3b3f271e3-scripts\") pod \"cinder-db-sync-9pvrg\" (UID: \"f09789c2-52ed-4321-95f8-02c3b3f271e3\") " pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.726857 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f09789c2-52ed-4321-95f8-02c3b3f271e3-config-data\") pod \"cinder-db-sync-9pvrg\" (UID: \"f09789c2-52ed-4321-95f8-02c3b3f271e3\") " pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.727140 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f09789c2-52ed-4321-95f8-02c3b3f271e3-db-sync-config-data\") pod \"cinder-db-sync-9pvrg\" (UID: \"f09789c2-52ed-4321-95f8-02c3b3f271e3\") " pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.730401 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcnsl\" (UniqueName: \"kubernetes.io/projected/f09789c2-52ed-4321-95f8-02c3b3f271e3-kube-api-access-hcnsl\") pod \"cinder-db-sync-9pvrg\" (UID: \"f09789c2-52ed-4321-95f8-02c3b3f271e3\") " pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" Jan 29 16:28:00 crc kubenswrapper[4714]: I0129 16:28:00.840603 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" Jan 29 16:28:01 crc kubenswrapper[4714]: I0129 16:28:01.277725 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-9pvrg"] Jan 29 16:28:01 crc kubenswrapper[4714]: I0129 16:28:01.882180 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" event={"ID":"f09789c2-52ed-4321-95f8-02c3b3f271e3","Type":"ContainerStarted","Data":"915808b45f9976849e0c7fd3de3f5846a7cc68f45e22deab7e1b486fb97c6590"} Jan 29 16:28:05 crc kubenswrapper[4714]: E0129 16:28:05.190414 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:28:12 crc kubenswrapper[4714]: I0129 16:28:12.694400 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:28:17 crc kubenswrapper[4714]: E0129 16:28:17.752574 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:28:20 crc kubenswrapper[4714]: I0129 16:28:20.059051 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" event={"ID":"f09789c2-52ed-4321-95f8-02c3b3f271e3","Type":"ContainerStarted","Data":"83eeaf58ca15604fd125219fc4be09b86cbb0308fb89b0438c4ada31625917b5"} Jan 29 16:28:20 crc kubenswrapper[4714]: I0129 16:28:20.084801 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" podStartSLOduration=2.552132126 podStartE2EDuration="20.0847782s" podCreationTimestamp="2026-01-29 16:28:00 +0000 UTC" firstStartedPulling="2026-01-29 16:28:01.288231275 +0000 UTC m=+1087.808732395" lastFinishedPulling="2026-01-29 16:28:18.820877329 +0000 UTC m=+1105.341378469" observedRunningTime="2026-01-29 16:28:20.079298367 +0000 UTC m=+1106.599799507" watchObservedRunningTime="2026-01-29 16:28:20.0847782 +0000 UTC m=+1106.605279330" Jan 29 16:28:27 crc kubenswrapper[4714]: I0129 16:28:27.112476 4714 generic.go:334] "Generic (PLEG): container finished" podID="f09789c2-52ed-4321-95f8-02c3b3f271e3" containerID="83eeaf58ca15604fd125219fc4be09b86cbb0308fb89b0438c4ada31625917b5" exitCode=0 Jan 29 16:28:27 crc kubenswrapper[4714]: I0129 16:28:27.112543 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" event={"ID":"f09789c2-52ed-4321-95f8-02c3b3f271e3","Type":"ContainerDied","Data":"83eeaf58ca15604fd125219fc4be09b86cbb0308fb89b0438c4ada31625917b5"} Jan 29 16:28:28 crc kubenswrapper[4714]: E0129 16:28:28.185980 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:28:28 crc kubenswrapper[4714]: I0129 16:28:28.379244 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" Jan 29 16:28:28 crc kubenswrapper[4714]: I0129 16:28:28.550997 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f09789c2-52ed-4321-95f8-02c3b3f271e3-etc-machine-id\") pod \"f09789c2-52ed-4321-95f8-02c3b3f271e3\" (UID: \"f09789c2-52ed-4321-95f8-02c3b3f271e3\") " Jan 29 16:28:28 crc kubenswrapper[4714]: I0129 16:28:28.551067 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f09789c2-52ed-4321-95f8-02c3b3f271e3-db-sync-config-data\") pod \"f09789c2-52ed-4321-95f8-02c3b3f271e3\" (UID: \"f09789c2-52ed-4321-95f8-02c3b3f271e3\") " Jan 29 16:28:28 crc kubenswrapper[4714]: I0129 16:28:28.551110 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f09789c2-52ed-4321-95f8-02c3b3f271e3-config-data\") pod \"f09789c2-52ed-4321-95f8-02c3b3f271e3\" (UID: \"f09789c2-52ed-4321-95f8-02c3b3f271e3\") " Jan 29 16:28:28 crc kubenswrapper[4714]: I0129 16:28:28.551170 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f09789c2-52ed-4321-95f8-02c3b3f271e3-scripts\") pod \"f09789c2-52ed-4321-95f8-02c3b3f271e3\" (UID: \"f09789c2-52ed-4321-95f8-02c3b3f271e3\") " Jan 29 16:28:28 crc kubenswrapper[4714]: I0129 16:28:28.551222 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f09789c2-52ed-4321-95f8-02c3b3f271e3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f09789c2-52ed-4321-95f8-02c3b3f271e3" (UID: "f09789c2-52ed-4321-95f8-02c3b3f271e3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:28:28 crc kubenswrapper[4714]: I0129 16:28:28.551262 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcnsl\" (UniqueName: \"kubernetes.io/projected/f09789c2-52ed-4321-95f8-02c3b3f271e3-kube-api-access-hcnsl\") pod \"f09789c2-52ed-4321-95f8-02c3b3f271e3\" (UID: \"f09789c2-52ed-4321-95f8-02c3b3f271e3\") " Jan 29 16:28:28 crc kubenswrapper[4714]: I0129 16:28:28.551682 4714 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f09789c2-52ed-4321-95f8-02c3b3f271e3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 29 16:28:28 crc kubenswrapper[4714]: I0129 16:28:28.559075 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f09789c2-52ed-4321-95f8-02c3b3f271e3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f09789c2-52ed-4321-95f8-02c3b3f271e3" (UID: "f09789c2-52ed-4321-95f8-02c3b3f271e3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:28:28 crc kubenswrapper[4714]: I0129 16:28:28.559197 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f09789c2-52ed-4321-95f8-02c3b3f271e3-scripts" (OuterVolumeSpecName: "scripts") pod "f09789c2-52ed-4321-95f8-02c3b3f271e3" (UID: "f09789c2-52ed-4321-95f8-02c3b3f271e3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:28:28 crc kubenswrapper[4714]: I0129 16:28:28.562379 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f09789c2-52ed-4321-95f8-02c3b3f271e3-kube-api-access-hcnsl" (OuterVolumeSpecName: "kube-api-access-hcnsl") pod "f09789c2-52ed-4321-95f8-02c3b3f271e3" (UID: "f09789c2-52ed-4321-95f8-02c3b3f271e3"). InnerVolumeSpecName "kube-api-access-hcnsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:28:28 crc kubenswrapper[4714]: I0129 16:28:28.594611 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f09789c2-52ed-4321-95f8-02c3b3f271e3-config-data" (OuterVolumeSpecName: "config-data") pod "f09789c2-52ed-4321-95f8-02c3b3f271e3" (UID: "f09789c2-52ed-4321-95f8-02c3b3f271e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:28:28 crc kubenswrapper[4714]: I0129 16:28:28.654321 4714 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f09789c2-52ed-4321-95f8-02c3b3f271e3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:28:28 crc kubenswrapper[4714]: I0129 16:28:28.654417 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f09789c2-52ed-4321-95f8-02c3b3f271e3-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:28:28 crc kubenswrapper[4714]: I0129 16:28:28.654457 4714 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f09789c2-52ed-4321-95f8-02c3b3f271e3-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:28:28 crc kubenswrapper[4714]: I0129 16:28:28.654481 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcnsl\" (UniqueName: \"kubernetes.io/projected/f09789c2-52ed-4321-95f8-02c3b3f271e3-kube-api-access-hcnsl\") on node \"crc\" DevicePath \"\"" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.132335 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" event={"ID":"f09789c2-52ed-4321-95f8-02c3b3f271e3","Type":"ContainerDied","Data":"915808b45f9976849e0c7fd3de3f5846a7cc68f45e22deab7e1b486fb97c6590"} Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.132745 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="915808b45f9976849e0c7fd3de3f5846a7cc68f45e22deab7e1b486fb97c6590" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.132694 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-9pvrg" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.431248 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Jan 29 16:28:29 crc kubenswrapper[4714]: E0129 16:28:29.431690 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f09789c2-52ed-4321-95f8-02c3b3f271e3" containerName="cinder-db-sync" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.431707 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="f09789c2-52ed-4321-95f8-02c3b3f271e3" containerName="cinder-db-sync" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.431835 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="f09789c2-52ed-4321-95f8-02c3b3f271e3" containerName="cinder-db-sync" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.432745 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.438999 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-cinder-dockercfg-xw2vq" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.439406 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scripts" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.439603 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scheduler-config-data" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.439822 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-config-data" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.448007 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.449165 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.453774 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-backup-config-data" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.454547 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.462858 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.572792 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.573808 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.577538 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-volume-volume1-config-data" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.578600 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-config-data\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.578641 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-lib-modules\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.578661 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-run\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.578677 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0390b29-ac12-4c76-a954-8c7236d81661-scripts\") pod \"cinder-scheduler-0\" (UID: \"c0390b29-ac12-4c76-a954-8c7236d81661\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.578692 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-scripts\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.578718 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.578746 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.578800 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-sys\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.578820 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rxqx\" (UniqueName: \"kubernetes.io/projected/c0390b29-ac12-4c76-a954-8c7236d81661-kube-api-access-6rxqx\") pod \"cinder-scheduler-0\" (UID: \"c0390b29-ac12-4c76-a954-8c7236d81661\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.578845 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.578859 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-dev\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.578875 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0390b29-ac12-4c76-a954-8c7236d81661-config-data\") pod \"cinder-scheduler-0\" (UID: \"c0390b29-ac12-4c76-a954-8c7236d81661\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.578895 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.578908 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-config-data-custom\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.578924 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5dd5\" (UniqueName: \"kubernetes.io/projected/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-kube-api-access-s5dd5\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.578975 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.579031 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c0390b29-ac12-4c76-a954-8c7236d81661-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c0390b29-ac12-4c76-a954-8c7236d81661\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.579065 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-etc-nvme\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.579085 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0390b29-ac12-4c76-a954-8c7236d81661-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c0390b29-ac12-4c76-a954-8c7236d81661\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.614664 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.679945 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5dd5\" (UniqueName: \"kubernetes.io/projected/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-kube-api-access-s5dd5\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.679999 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680036 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680054 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680070 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680186 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c0390b29-ac12-4c76-a954-8c7236d81661-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c0390b29-ac12-4c76-a954-8c7236d81661\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680205 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680230 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c0390b29-ac12-4c76-a954-8c7236d81661-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c0390b29-ac12-4c76-a954-8c7236d81661\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680297 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-etc-nvme\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680366 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0390b29-ac12-4c76-a954-8c7236d81661-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c0390b29-ac12-4c76-a954-8c7236d81661\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680375 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680399 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-etc-nvme\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680408 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680455 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680498 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-config-data\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680520 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-lib-modules\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680539 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-run\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680562 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0390b29-ac12-4c76-a954-8c7236d81661-scripts\") pod \"cinder-scheduler-0\" (UID: \"c0390b29-ac12-4c76-a954-8c7236d81661\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680581 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-scripts\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680594 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-lib-modules\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680604 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e154d80-4b79-4f74-809e-c1c274ed4063-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680899 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-run\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.680981 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.681046 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.681102 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e154d80-4b79-4f74-809e-c1c274ed4063-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.681140 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.681154 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-sys\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.681178 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rxqx\" (UniqueName: \"kubernetes.io/projected/c0390b29-ac12-4c76-a954-8c7236d81661-kube-api-access-6rxqx\") pod \"cinder-scheduler-0\" (UID: \"c0390b29-ac12-4c76-a954-8c7236d81661\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.681193 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-sys\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.681209 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbcbm\" (UniqueName: \"kubernetes.io/projected/2e154d80-4b79-4f74-809e-c1c274ed4063-kube-api-access-dbcbm\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.681232 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.681258 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.681273 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-dev\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.681289 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-run\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.681321 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e154d80-4b79-4f74-809e-c1c274ed4063-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.681328 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.681342 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0390b29-ac12-4c76-a954-8c7236d81661-config-data\") pod \"cinder-scheduler-0\" (UID: \"c0390b29-ac12-4c76-a954-8c7236d81661\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.681359 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.681376 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-dev\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.681531 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-dev\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.681555 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-sys\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.681595 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.681610 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-config-data-custom\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.682038 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.689619 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0390b29-ac12-4c76-a954-8c7236d81661-scripts\") pod \"cinder-scheduler-0\" (UID: \"c0390b29-ac12-4c76-a954-8c7236d81661\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.690219 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0390b29-ac12-4c76-a954-8c7236d81661-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c0390b29-ac12-4c76-a954-8c7236d81661\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.693554 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-config-data-custom\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.695924 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-config-data\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.700353 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-scripts\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.701873 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0390b29-ac12-4c76-a954-8c7236d81661-config-data\") pod \"cinder-scheduler-0\" (UID: \"c0390b29-ac12-4c76-a954-8c7236d81661\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.711679 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5dd5\" (UniqueName: \"kubernetes.io/projected/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-kube-api-access-s5dd5\") pod \"cinder-backup-0\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.714890 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rxqx\" (UniqueName: \"kubernetes.io/projected/c0390b29-ac12-4c76-a954-8c7236d81661-kube-api-access-6rxqx\") pod \"cinder-scheduler-0\" (UID: \"c0390b29-ac12-4c76-a954-8c7236d81661\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.720399 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.721738 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.726296 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-api-config-data" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.737260 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.757031 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.769251 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.785151 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e154d80-4b79-4f74-809e-c1c274ed4063-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.785197 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbcbm\" (UniqueName: \"kubernetes.io/projected/2e154d80-4b79-4f74-809e-c1c274ed4063-kube-api-access-dbcbm\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.785217 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.785234 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-run\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.785254 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e154d80-4b79-4f74-809e-c1c274ed4063-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.785273 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-dev\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.785286 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-sys\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.785311 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.785334 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.785351 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.785368 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.785391 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.785406 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.785436 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e154d80-4b79-4f74-809e-c1c274ed4063-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.785779 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.785826 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.785849 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.785880 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.785969 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-sys\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.786090 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-run\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.788082 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-dev\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.788145 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.788197 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.788227 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.791377 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e154d80-4b79-4f74-809e-c1c274ed4063-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.792155 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e154d80-4b79-4f74-809e-c1c274ed4063-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.793647 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e154d80-4b79-4f74-809e-c1c274ed4063-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.819543 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbcbm\" (UniqueName: \"kubernetes.io/projected/2e154d80-4b79-4f74-809e-c1c274ed4063-kube-api-access-dbcbm\") pod \"cinder-volume-volume1-0\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.886661 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-scripts\") pod \"cinder-api-0\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.886960 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-logs\") pod \"cinder-api-0\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.886997 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.887058 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.887286 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-796r4\" (UniqueName: \"kubernetes.io/projected/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-kube-api-access-796r4\") pod \"cinder-api-0\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.887482 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-config-data-custom\") pod \"cinder-api-0\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.887522 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-config-data\") pod \"cinder-api-0\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.988893 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-796r4\" (UniqueName: \"kubernetes.io/projected/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-kube-api-access-796r4\") pod \"cinder-api-0\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.989196 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-config-data-custom\") pod \"cinder-api-0\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.989217 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-config-data\") pod \"cinder-api-0\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.989246 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-scripts\") pod \"cinder-api-0\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.989283 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-logs\") pod \"cinder-api-0\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.989302 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.989358 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.990298 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-logs\") pod \"cinder-api-0\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.992967 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-config-data-custom\") pod \"cinder-api-0\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.993350 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-scripts\") pod \"cinder-api-0\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:29 crc kubenswrapper[4714]: I0129 16:28:29.994007 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-config-data\") pod \"cinder-api-0\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:30 crc kubenswrapper[4714]: I0129 16:28:30.005255 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-796r4\" (UniqueName: \"kubernetes.io/projected/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-kube-api-access-796r4\") pod \"cinder-api-0\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:30 crc kubenswrapper[4714]: I0129 16:28:30.058063 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:30 crc kubenswrapper[4714]: I0129 16:28:30.152049 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Jan 29 16:28:30 crc kubenswrapper[4714]: W0129 16:28:30.159985 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e154d80_4b79_4f74_809e_c1c274ed4063.slice/crio-aa9a0a1ce04cf5e776943ebe4b4ffc887cc7f0d7111f44e2e55f143d9edbcb9b WatchSource:0}: Error finding container aa9a0a1ce04cf5e776943ebe4b4ffc887cc7f0d7111f44e2e55f143d9edbcb9b: Status 404 returned error can't find the container with id aa9a0a1ce04cf5e776943ebe4b4ffc887cc7f0d7111f44e2e55f143d9edbcb9b Jan 29 16:28:30 crc kubenswrapper[4714]: W0129 16:28:30.254654 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0390b29_ac12_4c76_a954_8c7236d81661.slice/crio-a7e3257c239df083499879bab78d2e91169e0bd8d92a8c5fac288924f3619908 WatchSource:0}: Error finding container a7e3257c239df083499879bab78d2e91169e0bd8d92a8c5fac288924f3619908: Status 404 returned error can't find the container with id a7e3257c239df083499879bab78d2e91169e0bd8d92a8c5fac288924f3619908 Jan 29 16:28:30 crc kubenswrapper[4714]: I0129 16:28:30.255186 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Jan 29 16:28:30 crc kubenswrapper[4714]: W0129 16:28:30.262403 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode01c83c7_65ba_4f1b_9d17_ba5a824216bb.slice/crio-198a3df4efafc7e3421f9f093f38267c1562e37ed9290df232bcf8b82972d9a2 WatchSource:0}: Error finding container 198a3df4efafc7e3421f9f093f38267c1562e37ed9290df232bcf8b82972d9a2: Status 404 returned error can't find the container with id 198a3df4efafc7e3421f9f093f38267c1562e37ed9290df232bcf8b82972d9a2 Jan 29 16:28:30 crc kubenswrapper[4714]: I0129 16:28:30.262741 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Jan 29 16:28:30 crc kubenswrapper[4714]: I0129 16:28:30.500032 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Jan 29 16:28:30 crc kubenswrapper[4714]: W0129 16:28:30.506857 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b21fd86_5b9b_4b13_82aa_eb3d7f1fafb3.slice/crio-4d4f411f7f92df4f5d547c75bfc715baa406061177023b383b53537927256cc7 WatchSource:0}: Error finding container 4d4f411f7f92df4f5d547c75bfc715baa406061177023b383b53537927256cc7: Status 404 returned error can't find the container with id 4d4f411f7f92df4f5d547c75bfc715baa406061177023b383b53537927256cc7 Jan 29 16:28:31 crc kubenswrapper[4714]: I0129 16:28:31.148398 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"c0390b29-ac12-4c76-a954-8c7236d81661","Type":"ContainerStarted","Data":"a7e3257c239df083499879bab78d2e91169e0bd8d92a8c5fac288924f3619908"} Jan 29 16:28:31 crc kubenswrapper[4714]: I0129 16:28:31.150244 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3","Type":"ContainerStarted","Data":"79461c5a69e4029a68e7e8faa95d70a67ee30b1d8f5d9b8793dd7214e3da2c42"} Jan 29 16:28:31 crc kubenswrapper[4714]: I0129 16:28:31.150299 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3","Type":"ContainerStarted","Data":"4d4f411f7f92df4f5d547c75bfc715baa406061177023b383b53537927256cc7"} Jan 29 16:28:31 crc kubenswrapper[4714]: I0129 16:28:31.151489 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"2e154d80-4b79-4f74-809e-c1c274ed4063","Type":"ContainerStarted","Data":"aa9a0a1ce04cf5e776943ebe4b4ffc887cc7f0d7111f44e2e55f143d9edbcb9b"} Jan 29 16:28:31 crc kubenswrapper[4714]: I0129 16:28:31.152439 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"e01c83c7-65ba-4f1b-9d17-ba5a824216bb","Type":"ContainerStarted","Data":"198a3df4efafc7e3421f9f093f38267c1562e37ed9290df232bcf8b82972d9a2"} Jan 29 16:28:32 crc kubenswrapper[4714]: I0129 16:28:32.164561 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"c0390b29-ac12-4c76-a954-8c7236d81661","Type":"ContainerStarted","Data":"8b1dd6490cae9482eb69b08b1702b33a571bfe8525bbdf572a2f8f5feb7f4e29"} Jan 29 16:28:32 crc kubenswrapper[4714]: I0129 16:28:32.167708 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3","Type":"ContainerStarted","Data":"e1b4f5a218866bd462e639ef9e2c44453f1907feb94c13e36f29e0c48f866074"} Jan 29 16:28:32 crc kubenswrapper[4714]: I0129 16:28:32.167978 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:32 crc kubenswrapper[4714]: I0129 16:28:32.172208 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"2e154d80-4b79-4f74-809e-c1c274ed4063","Type":"ContainerStarted","Data":"fa59fe5c6301744749846298c687355a593a6ea20971c6048c481907a9c337d4"} Jan 29 16:28:32 crc kubenswrapper[4714]: I0129 16:28:32.172277 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"2e154d80-4b79-4f74-809e-c1c274ed4063","Type":"ContainerStarted","Data":"20125722bd9150df390f3154c0604d05230bfe6813ac735f8e571c5a8d6f0b17"} Jan 29 16:28:32 crc kubenswrapper[4714]: I0129 16:28:32.176250 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"e01c83c7-65ba-4f1b-9d17-ba5a824216bb","Type":"ContainerStarted","Data":"e1bfab0c04a18d8e21e126e8983546eaa8cd1254da1312d48182dae10e5944bc"} Jan 29 16:28:32 crc kubenswrapper[4714]: I0129 16:28:32.176298 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"e01c83c7-65ba-4f1b-9d17-ba5a824216bb","Type":"ContainerStarted","Data":"f457144c29e9469b286400011d4bf2e3e4f7a3f73d20a9a7750c293dc9d6911b"} Jan 29 16:28:32 crc kubenswrapper[4714]: I0129 16:28:32.197635 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-0" podStartSLOduration=3.197615522 podStartE2EDuration="3.197615522s" podCreationTimestamp="2026-01-29 16:28:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:28:32.189818478 +0000 UTC m=+1118.710319608" watchObservedRunningTime="2026-01-29 16:28:32.197615522 +0000 UTC m=+1118.718116652" Jan 29 16:28:32 crc kubenswrapper[4714]: I0129 16:28:32.223210 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-backup-0" podStartSLOduration=2.378593251 podStartE2EDuration="3.223193821s" podCreationTimestamp="2026-01-29 16:28:29 +0000 UTC" firstStartedPulling="2026-01-29 16:28:30.265461075 +0000 UTC m=+1116.785962195" lastFinishedPulling="2026-01-29 16:28:31.110061655 +0000 UTC m=+1117.630562765" observedRunningTime="2026-01-29 16:28:32.21971189 +0000 UTC m=+1118.740213010" watchObservedRunningTime="2026-01-29 16:28:32.223193821 +0000 UTC m=+1118.743694941" Jan 29 16:28:32 crc kubenswrapper[4714]: I0129 16:28:32.257096 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podStartSLOduration=2.23028182 podStartE2EDuration="3.257070297s" podCreationTimestamp="2026-01-29 16:28:29 +0000 UTC" firstStartedPulling="2026-01-29 16:28:30.161873254 +0000 UTC m=+1116.682374374" lastFinishedPulling="2026-01-29 16:28:31.188661731 +0000 UTC m=+1117.709162851" observedRunningTime="2026-01-29 16:28:32.248598966 +0000 UTC m=+1118.769100106" watchObservedRunningTime="2026-01-29 16:28:32.257070297 +0000 UTC m=+1118.777571417" Jan 29 16:28:33 crc kubenswrapper[4714]: I0129 16:28:33.185693 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"c0390b29-ac12-4c76-a954-8c7236d81661","Type":"ContainerStarted","Data":"a3a659f2d53d2c5e83b9c4601491f3c732c7a7d0fe867de849d8aab7c8e4a69c"} Jan 29 16:28:33 crc kubenswrapper[4714]: I0129 16:28:33.204131 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-scheduler-0" podStartSLOduration=3.3571337740000002 podStartE2EDuration="4.204111326s" podCreationTimestamp="2026-01-29 16:28:29 +0000 UTC" firstStartedPulling="2026-01-29 16:28:30.259593951 +0000 UTC m=+1116.780095061" lastFinishedPulling="2026-01-29 16:28:31.106571493 +0000 UTC m=+1117.627072613" observedRunningTime="2026-01-29 16:28:33.202657638 +0000 UTC m=+1119.723158758" watchObservedRunningTime="2026-01-29 16:28:33.204111326 +0000 UTC m=+1119.724612446" Jan 29 16:28:34 crc kubenswrapper[4714]: I0129 16:28:34.196057 4714 generic.go:334] "Generic (PLEG): container finished" podID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerID="fa59fe5c6301744749846298c687355a593a6ea20971c6048c481907a9c337d4" exitCode=1 Jan 29 16:28:34 crc kubenswrapper[4714]: I0129 16:28:34.196142 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"2e154d80-4b79-4f74-809e-c1c274ed4063","Type":"ContainerDied","Data":"fa59fe5c6301744749846298c687355a593a6ea20971c6048c481907a9c337d4"} Jan 29 16:28:34 crc kubenswrapper[4714]: I0129 16:28:34.197245 4714 scope.go:117] "RemoveContainer" containerID="fa59fe5c6301744749846298c687355a593a6ea20971c6048c481907a9c337d4" Jan 29 16:28:34 crc kubenswrapper[4714]: I0129 16:28:34.758250 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:28:34 crc kubenswrapper[4714]: I0129 16:28:34.769694 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:34 crc kubenswrapper[4714]: I0129 16:28:34.887542 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:35 crc kubenswrapper[4714]: I0129 16:28:35.204167 4714 generic.go:334] "Generic (PLEG): container finished" podID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerID="20125722bd9150df390f3154c0604d05230bfe6813ac735f8e571c5a8d6f0b17" exitCode=1 Jan 29 16:28:35 crc kubenswrapper[4714]: I0129 16:28:35.205077 4714 scope.go:117] "RemoveContainer" containerID="20125722bd9150df390f3154c0604d05230bfe6813ac735f8e571c5a8d6f0b17" Jan 29 16:28:35 crc kubenswrapper[4714]: I0129 16:28:35.205286 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"2e154d80-4b79-4f74-809e-c1c274ed4063","Type":"ContainerDied","Data":"20125722bd9150df390f3154c0604d05230bfe6813ac735f8e571c5a8d6f0b17"} Jan 29 16:28:35 crc kubenswrapper[4714]: I0129 16:28:35.205316 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"2e154d80-4b79-4f74-809e-c1c274ed4063","Type":"ContainerStarted","Data":"34f2b3f7b5334da41b022ccd7abc5bd0f409e91fd9b5a369d64fab6c7fb44523"} Jan 29 16:28:35 crc kubenswrapper[4714]: I0129 16:28:35.888139 4714 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:36 crc kubenswrapper[4714]: I0129 16:28:36.214741 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"2e154d80-4b79-4f74-809e-c1c274ed4063","Type":"ContainerStarted","Data":"46a4b26eb4dfed202b650ea94571e607a9817d63a8156ea95e8371d72e104ee7"} Jan 29 16:28:37 crc kubenswrapper[4714]: I0129 16:28:37.226095 4714 generic.go:334] "Generic (PLEG): container finished" podID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerID="34f2b3f7b5334da41b022ccd7abc5bd0f409e91fd9b5a369d64fab6c7fb44523" exitCode=1 Jan 29 16:28:37 crc kubenswrapper[4714]: I0129 16:28:37.226170 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"2e154d80-4b79-4f74-809e-c1c274ed4063","Type":"ContainerDied","Data":"34f2b3f7b5334da41b022ccd7abc5bd0f409e91fd9b5a369d64fab6c7fb44523"} Jan 29 16:28:37 crc kubenswrapper[4714]: I0129 16:28:37.226499 4714 scope.go:117] "RemoveContainer" containerID="fa59fe5c6301744749846298c687355a593a6ea20971c6048c481907a9c337d4" Jan 29 16:28:37 crc kubenswrapper[4714]: I0129 16:28:37.227106 4714 scope.go:117] "RemoveContainer" containerID="34f2b3f7b5334da41b022ccd7abc5bd0f409e91fd9b5a369d64fab6c7fb44523" Jan 29 16:28:37 crc kubenswrapper[4714]: E0129 16:28:37.227421 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\"" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" Jan 29 16:28:38 crc kubenswrapper[4714]: I0129 16:28:38.235572 4714 generic.go:334] "Generic (PLEG): container finished" podID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerID="46a4b26eb4dfed202b650ea94571e607a9817d63a8156ea95e8371d72e104ee7" exitCode=1 Jan 29 16:28:38 crc kubenswrapper[4714]: I0129 16:28:38.235615 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"2e154d80-4b79-4f74-809e-c1c274ed4063","Type":"ContainerDied","Data":"46a4b26eb4dfed202b650ea94571e607a9817d63a8156ea95e8371d72e104ee7"} Jan 29 16:28:38 crc kubenswrapper[4714]: I0129 16:28:38.235673 4714 scope.go:117] "RemoveContainer" containerID="20125722bd9150df390f3154c0604d05230bfe6813ac735f8e571c5a8d6f0b17" Jan 29 16:28:38 crc kubenswrapper[4714]: I0129 16:28:38.236225 4714 scope.go:117] "RemoveContainer" containerID="46a4b26eb4dfed202b650ea94571e607a9817d63a8156ea95e8371d72e104ee7" Jan 29 16:28:38 crc kubenswrapper[4714]: I0129 16:28:38.236265 4714 scope.go:117] "RemoveContainer" containerID="34f2b3f7b5334da41b022ccd7abc5bd0f409e91fd9b5a369d64fab6c7fb44523" Jan 29 16:28:38 crc kubenswrapper[4714]: E0129 16:28:38.236463 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" Jan 29 16:28:38 crc kubenswrapper[4714]: I0129 16:28:38.887514 4714 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:39 crc kubenswrapper[4714]: I0129 16:28:39.246695 4714 scope.go:117] "RemoveContainer" containerID="46a4b26eb4dfed202b650ea94571e607a9817d63a8156ea95e8371d72e104ee7" Jan 29 16:28:39 crc kubenswrapper[4714]: I0129 16:28:39.247284 4714 scope.go:117] "RemoveContainer" containerID="34f2b3f7b5334da41b022ccd7abc5bd0f409e91fd9b5a369d64fab6c7fb44523" Jan 29 16:28:39 crc kubenswrapper[4714]: E0129 16:28:39.247566 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" Jan 29 16:28:39 crc kubenswrapper[4714]: I0129 16:28:39.888030 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:39 crc kubenswrapper[4714]: I0129 16:28:39.888385 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:39 crc kubenswrapper[4714]: I0129 16:28:39.997319 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:28:40 crc kubenswrapper[4714]: I0129 16:28:40.009390 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:28:40 crc kubenswrapper[4714]: I0129 16:28:40.256047 4714 scope.go:117] "RemoveContainer" containerID="46a4b26eb4dfed202b650ea94571e607a9817d63a8156ea95e8371d72e104ee7" Jan 29 16:28:40 crc kubenswrapper[4714]: I0129 16:28:40.256079 4714 scope.go:117] "RemoveContainer" containerID="34f2b3f7b5334da41b022ccd7abc5bd0f409e91fd9b5a369d64fab6c7fb44523" Jan 29 16:28:40 crc kubenswrapper[4714]: E0129 16:28:40.256392 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" Jan 29 16:28:41 crc kubenswrapper[4714]: I0129 16:28:41.263118 4714 scope.go:117] "RemoveContainer" containerID="46a4b26eb4dfed202b650ea94571e607a9817d63a8156ea95e8371d72e104ee7" Jan 29 16:28:41 crc kubenswrapper[4714]: I0129 16:28:41.263148 4714 scope.go:117] "RemoveContainer" containerID="34f2b3f7b5334da41b022ccd7abc5bd0f409e91fd9b5a369d64fab6c7fb44523" Jan 29 16:28:41 crc kubenswrapper[4714]: E0129 16:28:41.263375 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" Jan 29 16:28:42 crc kubenswrapper[4714]: I0129 16:28:42.152547 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:28:42 crc kubenswrapper[4714]: E0129 16:28:42.196501 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:28:43 crc kubenswrapper[4714]: I0129 16:28:43.901571 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Jan 29 16:28:43 crc kubenswrapper[4714]: I0129 16:28:43.902882 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:28:43 crc kubenswrapper[4714]: I0129 16:28:43.910603 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Jan 29 16:28:44 crc kubenswrapper[4714]: I0129 16:28:44.045510 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8htn\" (UniqueName: \"kubernetes.io/projected/73e3160b-2922-47cf-999d-ad759cae98bc-kube-api-access-q8htn\") pod \"cinder-scheduler-1\" (UID: \"73e3160b-2922-47cf-999d-ad759cae98bc\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:28:44 crc kubenswrapper[4714]: I0129 16:28:44.045577 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73e3160b-2922-47cf-999d-ad759cae98bc-config-data\") pod \"cinder-scheduler-1\" (UID: \"73e3160b-2922-47cf-999d-ad759cae98bc\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:28:44 crc kubenswrapper[4714]: I0129 16:28:44.045697 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73e3160b-2922-47cf-999d-ad759cae98bc-config-data-custom\") pod \"cinder-scheduler-1\" (UID: \"73e3160b-2922-47cf-999d-ad759cae98bc\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:28:44 crc kubenswrapper[4714]: I0129 16:28:44.045728 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/73e3160b-2922-47cf-999d-ad759cae98bc-etc-machine-id\") pod \"cinder-scheduler-1\" (UID: \"73e3160b-2922-47cf-999d-ad759cae98bc\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:28:44 crc kubenswrapper[4714]: I0129 16:28:44.045769 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73e3160b-2922-47cf-999d-ad759cae98bc-scripts\") pod \"cinder-scheduler-1\" (UID: \"73e3160b-2922-47cf-999d-ad759cae98bc\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:28:44 crc kubenswrapper[4714]: I0129 16:28:44.147691 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73e3160b-2922-47cf-999d-ad759cae98bc-config-data-custom\") pod \"cinder-scheduler-1\" (UID: \"73e3160b-2922-47cf-999d-ad759cae98bc\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:28:44 crc kubenswrapper[4714]: I0129 16:28:44.147742 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/73e3160b-2922-47cf-999d-ad759cae98bc-etc-machine-id\") pod \"cinder-scheduler-1\" (UID: \"73e3160b-2922-47cf-999d-ad759cae98bc\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:28:44 crc kubenswrapper[4714]: I0129 16:28:44.147787 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73e3160b-2922-47cf-999d-ad759cae98bc-scripts\") pod \"cinder-scheduler-1\" (UID: \"73e3160b-2922-47cf-999d-ad759cae98bc\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:28:44 crc kubenswrapper[4714]: I0129 16:28:44.147845 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8htn\" (UniqueName: \"kubernetes.io/projected/73e3160b-2922-47cf-999d-ad759cae98bc-kube-api-access-q8htn\") pod \"cinder-scheduler-1\" (UID: \"73e3160b-2922-47cf-999d-ad759cae98bc\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:28:44 crc kubenswrapper[4714]: I0129 16:28:44.147883 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73e3160b-2922-47cf-999d-ad759cae98bc-config-data\") pod \"cinder-scheduler-1\" (UID: \"73e3160b-2922-47cf-999d-ad759cae98bc\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:28:44 crc kubenswrapper[4714]: I0129 16:28:44.147916 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/73e3160b-2922-47cf-999d-ad759cae98bc-etc-machine-id\") pod \"cinder-scheduler-1\" (UID: \"73e3160b-2922-47cf-999d-ad759cae98bc\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:28:44 crc kubenswrapper[4714]: I0129 16:28:44.158453 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73e3160b-2922-47cf-999d-ad759cae98bc-scripts\") pod \"cinder-scheduler-1\" (UID: \"73e3160b-2922-47cf-999d-ad759cae98bc\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:28:44 crc kubenswrapper[4714]: I0129 16:28:44.161004 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73e3160b-2922-47cf-999d-ad759cae98bc-config-data\") pod \"cinder-scheduler-1\" (UID: \"73e3160b-2922-47cf-999d-ad759cae98bc\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:28:44 crc kubenswrapper[4714]: I0129 16:28:44.163506 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73e3160b-2922-47cf-999d-ad759cae98bc-config-data-custom\") pod \"cinder-scheduler-1\" (UID: \"73e3160b-2922-47cf-999d-ad759cae98bc\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:28:44 crc kubenswrapper[4714]: I0129 16:28:44.168153 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8htn\" (UniqueName: \"kubernetes.io/projected/73e3160b-2922-47cf-999d-ad759cae98bc-kube-api-access-q8htn\") pod \"cinder-scheduler-1\" (UID: \"73e3160b-2922-47cf-999d-ad759cae98bc\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:28:44 crc kubenswrapper[4714]: I0129 16:28:44.255605 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:28:44 crc kubenswrapper[4714]: I0129 16:28:44.678644 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Jan 29 16:28:44 crc kubenswrapper[4714]: W0129 16:28:44.678920 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73e3160b_2922_47cf_999d_ad759cae98bc.slice/crio-e286b3bb2b844ed93f167bb8d414988c54c125f6caa42afdb169333e416d9923 WatchSource:0}: Error finding container e286b3bb2b844ed93f167bb8d414988c54c125f6caa42afdb169333e416d9923: Status 404 returned error can't find the container with id e286b3bb2b844ed93f167bb8d414988c54c125f6caa42afdb169333e416d9923 Jan 29 16:28:45 crc kubenswrapper[4714]: I0129 16:28:45.303528 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"73e3160b-2922-47cf-999d-ad759cae98bc","Type":"ContainerStarted","Data":"b78224b7eaf29fa5f35f5a14a7913fb0a21443ce839155a632681f4c3351fc14"} Jan 29 16:28:45 crc kubenswrapper[4714]: I0129 16:28:45.304953 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"73e3160b-2922-47cf-999d-ad759cae98bc","Type":"ContainerStarted","Data":"e286b3bb2b844ed93f167bb8d414988c54c125f6caa42afdb169333e416d9923"} Jan 29 16:28:46 crc kubenswrapper[4714]: I0129 16:28:46.325483 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"73e3160b-2922-47cf-999d-ad759cae98bc","Type":"ContainerStarted","Data":"887a921e7b634430b2e58452fdae19c42793bfa2ce2b387e0375f8cbedc76627"} Jan 29 16:28:49 crc kubenswrapper[4714]: I0129 16:28:49.256725 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:28:52 crc kubenswrapper[4714]: I0129 16:28:52.184372 4714 scope.go:117] "RemoveContainer" containerID="46a4b26eb4dfed202b650ea94571e607a9817d63a8156ea95e8371d72e104ee7" Jan 29 16:28:52 crc kubenswrapper[4714]: I0129 16:28:52.184681 4714 scope.go:117] "RemoveContainer" containerID="34f2b3f7b5334da41b022ccd7abc5bd0f409e91fd9b5a369d64fab6c7fb44523" Jan 29 16:28:53 crc kubenswrapper[4714]: I0129 16:28:53.394561 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"2e154d80-4b79-4f74-809e-c1c274ed4063","Type":"ContainerStarted","Data":"cdf72985bc60f6ce06d587a8a7eb4e0fd4beed9399bf382db190b6ec4763fc7c"} Jan 29 16:28:53 crc kubenswrapper[4714]: I0129 16:28:53.395070 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"2e154d80-4b79-4f74-809e-c1c274ed4063","Type":"ContainerStarted","Data":"f087542636afa3fd11613ddd38d7bd61a4a8cac1161d258a9bca9a6772f2da3c"} Jan 29 16:28:53 crc kubenswrapper[4714]: I0129 16:28:53.427635 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-scheduler-1" podStartSLOduration=10.427620308 podStartE2EDuration="10.427620308s" podCreationTimestamp="2026-01-29 16:28:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:28:46.35044233 +0000 UTC m=+1132.870943450" watchObservedRunningTime="2026-01-29 16:28:53.427620308 +0000 UTC m=+1139.948121428" Jan 29 16:28:54 crc kubenswrapper[4714]: E0129 16:28:54.330850 4714 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 16:28:54 crc kubenswrapper[4714]: E0129 16:28:54.331595 4714 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tmm88,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-wdwq5_openshift-marketplace(8c12ad14-f878-42a1-a168-bad4026ec2dd): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 16:28:54 crc kubenswrapper[4714]: E0129 16:28:54.333021 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.406624 4714 generic.go:334] "Generic (PLEG): container finished" podID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerID="cdf72985bc60f6ce06d587a8a7eb4e0fd4beed9399bf382db190b6ec4763fc7c" exitCode=1 Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.406665 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"2e154d80-4b79-4f74-809e-c1c274ed4063","Type":"ContainerDied","Data":"cdf72985bc60f6ce06d587a8a7eb4e0fd4beed9399bf382db190b6ec4763fc7c"} Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.406699 4714 scope.go:117] "RemoveContainer" containerID="34f2b3f7b5334da41b022ccd7abc5bd0f409e91fd9b5a369d64fab6c7fb44523" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.407385 4714 scope.go:117] "RemoveContainer" containerID="cdf72985bc60f6ce06d587a8a7eb4e0fd4beed9399bf382db190b6ec4763fc7c" Jan 29 16:28:54 crc kubenswrapper[4714]: E0129 16:28:54.407723 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\"" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.444594 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.496351 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.497460 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.514705 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.620674 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67fa6a5a-7bfb-4079-8658-aca62c22bc73-scripts\") pod \"cinder-scheduler-2\" (UID: \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.620718 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67fa6a5a-7bfb-4079-8658-aca62c22bc73-config-data-custom\") pod \"cinder-scheduler-2\" (UID: \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.620738 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mf9c\" (UniqueName: \"kubernetes.io/projected/67fa6a5a-7bfb-4079-8658-aca62c22bc73-kube-api-access-2mf9c\") pod \"cinder-scheduler-2\" (UID: \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.620762 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67fa6a5a-7bfb-4079-8658-aca62c22bc73-config-data\") pod \"cinder-scheduler-2\" (UID: \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.620800 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67fa6a5a-7bfb-4079-8658-aca62c22bc73-etc-machine-id\") pod \"cinder-scheduler-2\" (UID: \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.723068 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67fa6a5a-7bfb-4079-8658-aca62c22bc73-scripts\") pod \"cinder-scheduler-2\" (UID: \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.723142 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67fa6a5a-7bfb-4079-8658-aca62c22bc73-config-data-custom\") pod \"cinder-scheduler-2\" (UID: \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.723175 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mf9c\" (UniqueName: \"kubernetes.io/projected/67fa6a5a-7bfb-4079-8658-aca62c22bc73-kube-api-access-2mf9c\") pod \"cinder-scheduler-2\" (UID: \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.723220 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67fa6a5a-7bfb-4079-8658-aca62c22bc73-config-data\") pod \"cinder-scheduler-2\" (UID: \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.723258 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67fa6a5a-7bfb-4079-8658-aca62c22bc73-etc-machine-id\") pod \"cinder-scheduler-2\" (UID: \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.723416 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67fa6a5a-7bfb-4079-8658-aca62c22bc73-etc-machine-id\") pod \"cinder-scheduler-2\" (UID: \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.731364 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67fa6a5a-7bfb-4079-8658-aca62c22bc73-config-data-custom\") pod \"cinder-scheduler-2\" (UID: \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.733288 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67fa6a5a-7bfb-4079-8658-aca62c22bc73-config-data\") pod \"cinder-scheduler-2\" (UID: \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.735270 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67fa6a5a-7bfb-4079-8658-aca62c22bc73-scripts\") pod \"cinder-scheduler-2\" (UID: \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.749858 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mf9c\" (UniqueName: \"kubernetes.io/projected/67fa6a5a-7bfb-4079-8658-aca62c22bc73-kube-api-access-2mf9c\") pod \"cinder-scheduler-2\" (UID: \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.815686 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:28:54 crc kubenswrapper[4714]: I0129 16:28:54.887857 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:55 crc kubenswrapper[4714]: I0129 16:28:55.275083 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Jan 29 16:28:55 crc kubenswrapper[4714]: I0129 16:28:55.421671 4714 scope.go:117] "RemoveContainer" containerID="cdf72985bc60f6ce06d587a8a7eb4e0fd4beed9399bf382db190b6ec4763fc7c" Jan 29 16:28:55 crc kubenswrapper[4714]: E0129 16:28:55.422090 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\"" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" Jan 29 16:28:55 crc kubenswrapper[4714]: I0129 16:28:55.427113 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"67fa6a5a-7bfb-4079-8658-aca62c22bc73","Type":"ContainerStarted","Data":"41540f6e31a993ff1d30e0d04c9d92278e10fe8c05007eecaef75b03ade05470"} Jan 29 16:28:56 crc kubenswrapper[4714]: I0129 16:28:56.439553 4714 generic.go:334] "Generic (PLEG): container finished" podID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerID="f087542636afa3fd11613ddd38d7bd61a4a8cac1161d258a9bca9a6772f2da3c" exitCode=1 Jan 29 16:28:56 crc kubenswrapper[4714]: I0129 16:28:56.439676 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"2e154d80-4b79-4f74-809e-c1c274ed4063","Type":"ContainerDied","Data":"f087542636afa3fd11613ddd38d7bd61a4a8cac1161d258a9bca9a6772f2da3c"} Jan 29 16:28:56 crc kubenswrapper[4714]: I0129 16:28:56.439734 4714 scope.go:117] "RemoveContainer" containerID="46a4b26eb4dfed202b650ea94571e607a9817d63a8156ea95e8371d72e104ee7" Jan 29 16:28:56 crc kubenswrapper[4714]: I0129 16:28:56.441995 4714 scope.go:117] "RemoveContainer" containerID="f087542636afa3fd11613ddd38d7bd61a4a8cac1161d258a9bca9a6772f2da3c" Jan 29 16:28:56 crc kubenswrapper[4714]: I0129 16:28:56.442427 4714 scope.go:117] "RemoveContainer" containerID="cdf72985bc60f6ce06d587a8a7eb4e0fd4beed9399bf382db190b6ec4763fc7c" Jan 29 16:28:56 crc kubenswrapper[4714]: E0129 16:28:56.443194 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" Jan 29 16:28:56 crc kubenswrapper[4714]: I0129 16:28:56.455009 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"67fa6a5a-7bfb-4079-8658-aca62c22bc73","Type":"ContainerStarted","Data":"a901ff5206c34331366bd833f8b4176744c61709983ca47575647f7a5aef1760"} Jan 29 16:28:56 crc kubenswrapper[4714]: I0129 16:28:56.455063 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"67fa6a5a-7bfb-4079-8658-aca62c22bc73","Type":"ContainerStarted","Data":"a781996ba3c27232811a12f89e03f81db23d935d200f9ce52272f5303e895f04"} Jan 29 16:28:56 crc kubenswrapper[4714]: I0129 16:28:56.510128 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-scheduler-2" podStartSLOduration=2.510108933 podStartE2EDuration="2.510108933s" podCreationTimestamp="2026-01-29 16:28:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:28:56.504422725 +0000 UTC m=+1143.024923855" watchObservedRunningTime="2026-01-29 16:28:56.510108933 +0000 UTC m=+1143.030610043" Jan 29 16:28:56 crc kubenswrapper[4714]: I0129 16:28:56.887347 4714 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:57 crc kubenswrapper[4714]: I0129 16:28:57.478588 4714 scope.go:117] "RemoveContainer" containerID="f087542636afa3fd11613ddd38d7bd61a4a8cac1161d258a9bca9a6772f2da3c" Jan 29 16:28:57 crc kubenswrapper[4714]: I0129 16:28:57.478638 4714 scope.go:117] "RemoveContainer" containerID="cdf72985bc60f6ce06d587a8a7eb4e0fd4beed9399bf382db190b6ec4763fc7c" Jan 29 16:28:57 crc kubenswrapper[4714]: E0129 16:28:57.479263 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" Jan 29 16:28:59 crc kubenswrapper[4714]: I0129 16:28:59.816128 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:28:59 crc kubenswrapper[4714]: I0129 16:28:59.887913 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:28:59 crc kubenswrapper[4714]: I0129 16:28:59.888640 4714 scope.go:117] "RemoveContainer" containerID="f087542636afa3fd11613ddd38d7bd61a4a8cac1161d258a9bca9a6772f2da3c" Jan 29 16:28:59 crc kubenswrapper[4714]: I0129 16:28:59.888657 4714 scope.go:117] "RemoveContainer" containerID="cdf72985bc60f6ce06d587a8a7eb4e0fd4beed9399bf382db190b6ec4763fc7c" Jan 29 16:28:59 crc kubenswrapper[4714]: E0129 16:28:59.888856 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" Jan 29 16:29:05 crc kubenswrapper[4714]: I0129 16:29:05.060071 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:29:05 crc kubenswrapper[4714]: I0129 16:29:05.595254 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Jan 29 16:29:05 crc kubenswrapper[4714]: I0129 16:29:05.595499 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-2" podUID="67fa6a5a-7bfb-4079-8658-aca62c22bc73" containerName="cinder-scheduler" containerID="cri-o://a781996ba3c27232811a12f89e03f81db23d935d200f9ce52272f5303e895f04" gracePeriod=30 Jan 29 16:29:05 crc kubenswrapper[4714]: I0129 16:29:05.595735 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-2" podUID="67fa6a5a-7bfb-4079-8658-aca62c22bc73" containerName="probe" containerID="cri-o://a901ff5206c34331366bd833f8b4176744c61709983ca47575647f7a5aef1760" gracePeriod=30 Jan 29 16:29:06 crc kubenswrapper[4714]: E0129 16:29:06.187039 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:29:06 crc kubenswrapper[4714]: I0129 16:29:06.540646 4714 generic.go:334] "Generic (PLEG): container finished" podID="67fa6a5a-7bfb-4079-8658-aca62c22bc73" containerID="a901ff5206c34331366bd833f8b4176744c61709983ca47575647f7a5aef1760" exitCode=0 Jan 29 16:29:06 crc kubenswrapper[4714]: I0129 16:29:06.540713 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"67fa6a5a-7bfb-4079-8658-aca62c22bc73","Type":"ContainerDied","Data":"a901ff5206c34331366bd833f8b4176744c61709983ca47575647f7a5aef1760"} Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.028735 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.129504 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67fa6a5a-7bfb-4079-8658-aca62c22bc73-config-data-custom\") pod \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\" (UID: \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\") " Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.129629 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67fa6a5a-7bfb-4079-8658-aca62c22bc73-config-data\") pod \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\" (UID: \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\") " Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.129734 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mf9c\" (UniqueName: \"kubernetes.io/projected/67fa6a5a-7bfb-4079-8658-aca62c22bc73-kube-api-access-2mf9c\") pod \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\" (UID: \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\") " Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.129772 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67fa6a5a-7bfb-4079-8658-aca62c22bc73-etc-machine-id\") pod \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\" (UID: \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\") " Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.129804 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67fa6a5a-7bfb-4079-8658-aca62c22bc73-scripts\") pod \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\" (UID: \"67fa6a5a-7bfb-4079-8658-aca62c22bc73\") " Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.129966 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/67fa6a5a-7bfb-4079-8658-aca62c22bc73-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "67fa6a5a-7bfb-4079-8658-aca62c22bc73" (UID: "67fa6a5a-7bfb-4079-8658-aca62c22bc73"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.130583 4714 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67fa6a5a-7bfb-4079-8658-aca62c22bc73-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.136151 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67fa6a5a-7bfb-4079-8658-aca62c22bc73-scripts" (OuterVolumeSpecName: "scripts") pod "67fa6a5a-7bfb-4079-8658-aca62c22bc73" (UID: "67fa6a5a-7bfb-4079-8658-aca62c22bc73"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.136433 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67fa6a5a-7bfb-4079-8658-aca62c22bc73-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "67fa6a5a-7bfb-4079-8658-aca62c22bc73" (UID: "67fa6a5a-7bfb-4079-8658-aca62c22bc73"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.147303 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67fa6a5a-7bfb-4079-8658-aca62c22bc73-kube-api-access-2mf9c" (OuterVolumeSpecName: "kube-api-access-2mf9c") pod "67fa6a5a-7bfb-4079-8658-aca62c22bc73" (UID: "67fa6a5a-7bfb-4079-8658-aca62c22bc73"). InnerVolumeSpecName "kube-api-access-2mf9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.229006 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67fa6a5a-7bfb-4079-8658-aca62c22bc73-config-data" (OuterVolumeSpecName: "config-data") pod "67fa6a5a-7bfb-4079-8658-aca62c22bc73" (UID: "67fa6a5a-7bfb-4079-8658-aca62c22bc73"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.231191 4714 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67fa6a5a-7bfb-4079-8658-aca62c22bc73-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.231221 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67fa6a5a-7bfb-4079-8658-aca62c22bc73-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.231233 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67fa6a5a-7bfb-4079-8658-aca62c22bc73-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.231244 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mf9c\" (UniqueName: \"kubernetes.io/projected/67fa6a5a-7bfb-4079-8658-aca62c22bc73-kube-api-access-2mf9c\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.551298 4714 generic.go:334] "Generic (PLEG): container finished" podID="67fa6a5a-7bfb-4079-8658-aca62c22bc73" containerID="a781996ba3c27232811a12f89e03f81db23d935d200f9ce52272f5303e895f04" exitCode=0 Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.551348 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"67fa6a5a-7bfb-4079-8658-aca62c22bc73","Type":"ContainerDied","Data":"a781996ba3c27232811a12f89e03f81db23d935d200f9ce52272f5303e895f04"} Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.551358 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-2" Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.551379 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"67fa6a5a-7bfb-4079-8658-aca62c22bc73","Type":"ContainerDied","Data":"41540f6e31a993ff1d30e0d04c9d92278e10fe8c05007eecaef75b03ade05470"} Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.551403 4714 scope.go:117] "RemoveContainer" containerID="a901ff5206c34331366bd833f8b4176744c61709983ca47575647f7a5aef1760" Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.578917 4714 scope.go:117] "RemoveContainer" containerID="a781996ba3c27232811a12f89e03f81db23d935d200f9ce52272f5303e895f04" Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.596153 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.612198 4714 scope.go:117] "RemoveContainer" containerID="a901ff5206c34331366bd833f8b4176744c61709983ca47575647f7a5aef1760" Jan 29 16:29:07 crc kubenswrapper[4714]: E0129 16:29:07.613422 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a901ff5206c34331366bd833f8b4176744c61709983ca47575647f7a5aef1760\": container with ID starting with a901ff5206c34331366bd833f8b4176744c61709983ca47575647f7a5aef1760 not found: ID does not exist" containerID="a901ff5206c34331366bd833f8b4176744c61709983ca47575647f7a5aef1760" Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.613510 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a901ff5206c34331366bd833f8b4176744c61709983ca47575647f7a5aef1760"} err="failed to get container status \"a901ff5206c34331366bd833f8b4176744c61709983ca47575647f7a5aef1760\": rpc error: code = NotFound desc = could not find container \"a901ff5206c34331366bd833f8b4176744c61709983ca47575647f7a5aef1760\": container with ID starting with a901ff5206c34331366bd833f8b4176744c61709983ca47575647f7a5aef1760 not found: ID does not exist" Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.613543 4714 scope.go:117] "RemoveContainer" containerID="a781996ba3c27232811a12f89e03f81db23d935d200f9ce52272f5303e895f04" Jan 29 16:29:07 crc kubenswrapper[4714]: E0129 16:29:07.613888 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a781996ba3c27232811a12f89e03f81db23d935d200f9ce52272f5303e895f04\": container with ID starting with a781996ba3c27232811a12f89e03f81db23d935d200f9ce52272f5303e895f04 not found: ID does not exist" containerID="a781996ba3c27232811a12f89e03f81db23d935d200f9ce52272f5303e895f04" Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.613924 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a781996ba3c27232811a12f89e03f81db23d935d200f9ce52272f5303e895f04"} err="failed to get container status \"a781996ba3c27232811a12f89e03f81db23d935d200f9ce52272f5303e895f04\": rpc error: code = NotFound desc = could not find container \"a781996ba3c27232811a12f89e03f81db23d935d200f9ce52272f5303e895f04\": container with ID starting with a781996ba3c27232811a12f89e03f81db23d935d200f9ce52272f5303e895f04 not found: ID does not exist" Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.626308 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.652545 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.652768 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-1" podUID="73e3160b-2922-47cf-999d-ad759cae98bc" containerName="cinder-scheduler" containerID="cri-o://b78224b7eaf29fa5f35f5a14a7913fb0a21443ce839155a632681f4c3351fc14" gracePeriod=30 Jan 29 16:29:07 crc kubenswrapper[4714]: I0129 16:29:07.653119 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-1" podUID="73e3160b-2922-47cf-999d-ad759cae98bc" containerName="probe" containerID="cri-o://887a921e7b634430b2e58452fdae19c42793bfa2ce2b387e0375f8cbedc76627" gracePeriod=30 Jan 29 16:29:08 crc kubenswrapper[4714]: I0129 16:29:08.199903 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67fa6a5a-7bfb-4079-8658-aca62c22bc73" path="/var/lib/kubelet/pods/67fa6a5a-7bfb-4079-8658-aca62c22bc73/volumes" Jan 29 16:29:08 crc kubenswrapper[4714]: I0129 16:29:08.561011 4714 generic.go:334] "Generic (PLEG): container finished" podID="73e3160b-2922-47cf-999d-ad759cae98bc" containerID="887a921e7b634430b2e58452fdae19c42793bfa2ce2b387e0375f8cbedc76627" exitCode=0 Jan 29 16:29:08 crc kubenswrapper[4714]: I0129 16:29:08.561101 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"73e3160b-2922-47cf-999d-ad759cae98bc","Type":"ContainerDied","Data":"887a921e7b634430b2e58452fdae19c42793bfa2ce2b387e0375f8cbedc76627"} Jan 29 16:29:11 crc kubenswrapper[4714]: I0129 16:29:11.184362 4714 scope.go:117] "RemoveContainer" containerID="f087542636afa3fd11613ddd38d7bd61a4a8cac1161d258a9bca9a6772f2da3c" Jan 29 16:29:11 crc kubenswrapper[4714]: I0129 16:29:11.184668 4714 scope.go:117] "RemoveContainer" containerID="cdf72985bc60f6ce06d587a8a7eb4e0fd4beed9399bf382db190b6ec4763fc7c" Jan 29 16:29:11 crc kubenswrapper[4714]: E0129 16:29:11.185070 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.103316 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.210827 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73e3160b-2922-47cf-999d-ad759cae98bc-scripts\") pod \"73e3160b-2922-47cf-999d-ad759cae98bc\" (UID: \"73e3160b-2922-47cf-999d-ad759cae98bc\") " Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.210927 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/73e3160b-2922-47cf-999d-ad759cae98bc-etc-machine-id\") pod \"73e3160b-2922-47cf-999d-ad759cae98bc\" (UID: \"73e3160b-2922-47cf-999d-ad759cae98bc\") " Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.211099 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8htn\" (UniqueName: \"kubernetes.io/projected/73e3160b-2922-47cf-999d-ad759cae98bc-kube-api-access-q8htn\") pod \"73e3160b-2922-47cf-999d-ad759cae98bc\" (UID: \"73e3160b-2922-47cf-999d-ad759cae98bc\") " Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.211139 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73e3160b-2922-47cf-999d-ad759cae98bc-config-data\") pod \"73e3160b-2922-47cf-999d-ad759cae98bc\" (UID: \"73e3160b-2922-47cf-999d-ad759cae98bc\") " Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.211162 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73e3160b-2922-47cf-999d-ad759cae98bc-config-data-custom\") pod \"73e3160b-2922-47cf-999d-ad759cae98bc\" (UID: \"73e3160b-2922-47cf-999d-ad759cae98bc\") " Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.211159 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/73e3160b-2922-47cf-999d-ad759cae98bc-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "73e3160b-2922-47cf-999d-ad759cae98bc" (UID: "73e3160b-2922-47cf-999d-ad759cae98bc"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.211544 4714 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/73e3160b-2922-47cf-999d-ad759cae98bc-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.217113 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73e3160b-2922-47cf-999d-ad759cae98bc-scripts" (OuterVolumeSpecName: "scripts") pod "73e3160b-2922-47cf-999d-ad759cae98bc" (UID: "73e3160b-2922-47cf-999d-ad759cae98bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.219275 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73e3160b-2922-47cf-999d-ad759cae98bc-kube-api-access-q8htn" (OuterVolumeSpecName: "kube-api-access-q8htn") pod "73e3160b-2922-47cf-999d-ad759cae98bc" (UID: "73e3160b-2922-47cf-999d-ad759cae98bc"). InnerVolumeSpecName "kube-api-access-q8htn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.223148 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73e3160b-2922-47cf-999d-ad759cae98bc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "73e3160b-2922-47cf-999d-ad759cae98bc" (UID: "73e3160b-2922-47cf-999d-ad759cae98bc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.312639 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8htn\" (UniqueName: \"kubernetes.io/projected/73e3160b-2922-47cf-999d-ad759cae98bc-kube-api-access-q8htn\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.312666 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73e3160b-2922-47cf-999d-ad759cae98bc-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.312676 4714 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73e3160b-2922-47cf-999d-ad759cae98bc-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.313591 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73e3160b-2922-47cf-999d-ad759cae98bc-config-data" (OuterVolumeSpecName: "config-data") pod "73e3160b-2922-47cf-999d-ad759cae98bc" (UID: "73e3160b-2922-47cf-999d-ad759cae98bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.414408 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73e3160b-2922-47cf-999d-ad759cae98bc-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.596715 4714 generic.go:334] "Generic (PLEG): container finished" podID="73e3160b-2922-47cf-999d-ad759cae98bc" containerID="b78224b7eaf29fa5f35f5a14a7913fb0a21443ce839155a632681f4c3351fc14" exitCode=0 Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.596765 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"73e3160b-2922-47cf-999d-ad759cae98bc","Type":"ContainerDied","Data":"b78224b7eaf29fa5f35f5a14a7913fb0a21443ce839155a632681f4c3351fc14"} Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.596797 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-1" Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.596799 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"73e3160b-2922-47cf-999d-ad759cae98bc","Type":"ContainerDied","Data":"e286b3bb2b844ed93f167bb8d414988c54c125f6caa42afdb169333e416d9923"} Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.596819 4714 scope.go:117] "RemoveContainer" containerID="887a921e7b634430b2e58452fdae19c42793bfa2ce2b387e0375f8cbedc76627" Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.623316 4714 scope.go:117] "RemoveContainer" containerID="b78224b7eaf29fa5f35f5a14a7913fb0a21443ce839155a632681f4c3351fc14" Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.643057 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.647449 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.654806 4714 scope.go:117] "RemoveContainer" containerID="887a921e7b634430b2e58452fdae19c42793bfa2ce2b387e0375f8cbedc76627" Jan 29 16:29:12 crc kubenswrapper[4714]: E0129 16:29:12.655334 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"887a921e7b634430b2e58452fdae19c42793bfa2ce2b387e0375f8cbedc76627\": container with ID starting with 887a921e7b634430b2e58452fdae19c42793bfa2ce2b387e0375f8cbedc76627 not found: ID does not exist" containerID="887a921e7b634430b2e58452fdae19c42793bfa2ce2b387e0375f8cbedc76627" Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.655402 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"887a921e7b634430b2e58452fdae19c42793bfa2ce2b387e0375f8cbedc76627"} err="failed to get container status \"887a921e7b634430b2e58452fdae19c42793bfa2ce2b387e0375f8cbedc76627\": rpc error: code = NotFound desc = could not find container \"887a921e7b634430b2e58452fdae19c42793bfa2ce2b387e0375f8cbedc76627\": container with ID starting with 887a921e7b634430b2e58452fdae19c42793bfa2ce2b387e0375f8cbedc76627 not found: ID does not exist" Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.655447 4714 scope.go:117] "RemoveContainer" containerID="b78224b7eaf29fa5f35f5a14a7913fb0a21443ce839155a632681f4c3351fc14" Jan 29 16:29:12 crc kubenswrapper[4714]: E0129 16:29:12.659448 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b78224b7eaf29fa5f35f5a14a7913fb0a21443ce839155a632681f4c3351fc14\": container with ID starting with b78224b7eaf29fa5f35f5a14a7913fb0a21443ce839155a632681f4c3351fc14 not found: ID does not exist" containerID="b78224b7eaf29fa5f35f5a14a7913fb0a21443ce839155a632681f4c3351fc14" Jan 29 16:29:12 crc kubenswrapper[4714]: I0129 16:29:12.659506 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b78224b7eaf29fa5f35f5a14a7913fb0a21443ce839155a632681f4c3351fc14"} err="failed to get container status \"b78224b7eaf29fa5f35f5a14a7913fb0a21443ce839155a632681f4c3351fc14\": rpc error: code = NotFound desc = could not find container \"b78224b7eaf29fa5f35f5a14a7913fb0a21443ce839155a632681f4c3351fc14\": container with ID starting with b78224b7eaf29fa5f35f5a14a7913fb0a21443ce839155a632681f4c3351fc14 not found: ID does not exist" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.897078 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Jan 29 16:29:13 crc kubenswrapper[4714]: E0129 16:29:13.897626 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73e3160b-2922-47cf-999d-ad759cae98bc" containerName="cinder-scheduler" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.897641 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="73e3160b-2922-47cf-999d-ad759cae98bc" containerName="cinder-scheduler" Jan 29 16:29:13 crc kubenswrapper[4714]: E0129 16:29:13.897655 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67fa6a5a-7bfb-4079-8658-aca62c22bc73" containerName="cinder-scheduler" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.897663 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="67fa6a5a-7bfb-4079-8658-aca62c22bc73" containerName="cinder-scheduler" Jan 29 16:29:13 crc kubenswrapper[4714]: E0129 16:29:13.897686 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67fa6a5a-7bfb-4079-8658-aca62c22bc73" containerName="probe" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.897693 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="67fa6a5a-7bfb-4079-8658-aca62c22bc73" containerName="probe" Jan 29 16:29:13 crc kubenswrapper[4714]: E0129 16:29:13.897710 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73e3160b-2922-47cf-999d-ad759cae98bc" containerName="probe" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.897718 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="73e3160b-2922-47cf-999d-ad759cae98bc" containerName="probe" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.897863 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="73e3160b-2922-47cf-999d-ad759cae98bc" containerName="probe" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.897878 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="67fa6a5a-7bfb-4079-8658-aca62c22bc73" containerName="probe" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.897892 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="73e3160b-2922-47cf-999d-ad759cae98bc" containerName="cinder-scheduler" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.897902 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="67fa6a5a-7bfb-4079-8658-aca62c22bc73" containerName="cinder-scheduler" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.898742 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.907973 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.939442 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-etc-iscsi\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.939485 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-dev\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.939502 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-etc-machine-id\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.939519 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1df413aa-ba11-47aa-9b37-989956046c9f-config-data\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.939558 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-var-locks-cinder\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.939577 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-sys\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.939600 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1df413aa-ba11-47aa-9b37-989956046c9f-scripts\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.939662 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-var-lib-cinder\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.939687 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-etc-nvme\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.939706 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-var-locks-brick\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.939723 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmcjf\" (UniqueName: \"kubernetes.io/projected/1df413aa-ba11-47aa-9b37-989956046c9f-kube-api-access-kmcjf\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.939744 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-lib-modules\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.939773 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1df413aa-ba11-47aa-9b37-989956046c9f-config-data-custom\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:13 crc kubenswrapper[4714]: I0129 16:29:13.939802 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-run\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.040654 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-var-locks-cinder\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.040712 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-sys\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.040756 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1df413aa-ba11-47aa-9b37-989956046c9f-scripts\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.040807 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-var-lib-cinder\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.040831 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-etc-nvme\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.040863 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-var-locks-brick\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.040909 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-var-locks-cinder\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.040992 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-etc-nvme\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.040990 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-sys\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.041044 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-var-locks-brick\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.041063 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmcjf\" (UniqueName: \"kubernetes.io/projected/1df413aa-ba11-47aa-9b37-989956046c9f-kube-api-access-kmcjf\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.041050 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-var-lib-cinder\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.041151 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-lib-modules\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.041201 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1df413aa-ba11-47aa-9b37-989956046c9f-config-data-custom\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.041253 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-run\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.041327 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-lib-modules\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.041454 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-run\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.041870 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-etc-iscsi\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.041920 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-etc-iscsi\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.041963 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-dev\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.042007 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-etc-machine-id\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.042047 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-dev\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.042055 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1df413aa-ba11-47aa-9b37-989956046c9f-config-data\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.042076 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-etc-machine-id\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.047063 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1df413aa-ba11-47aa-9b37-989956046c9f-scripts\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.048350 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1df413aa-ba11-47aa-9b37-989956046c9f-config-data-custom\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.048739 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1df413aa-ba11-47aa-9b37-989956046c9f-config-data\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.064388 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmcjf\" (UniqueName: \"kubernetes.io/projected/1df413aa-ba11-47aa-9b37-989956046c9f-kube-api-access-kmcjf\") pod \"cinder-backup-1\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.199063 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73e3160b-2922-47cf-999d-ad759cae98bc" path="/var/lib/kubelet/pods/73e3160b-2922-47cf-999d-ad759cae98bc/volumes" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.213680 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:14 crc kubenswrapper[4714]: I0129 16:29:14.660047 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Jan 29 16:29:15 crc kubenswrapper[4714]: I0129 16:29:15.630530 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"1df413aa-ba11-47aa-9b37-989956046c9f","Type":"ContainerStarted","Data":"216659e8b1aeed2e1c0a3e238db1d9a8366df5059f31bffd8d1c1116d3eda51f"} Jan 29 16:29:15 crc kubenswrapper[4714]: I0129 16:29:15.631395 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"1df413aa-ba11-47aa-9b37-989956046c9f","Type":"ContainerStarted","Data":"11899e2f2c2124b1f13dd44e2372c23241e6ddb29215a4c11ae4fe8ed891c9bb"} Jan 29 16:29:15 crc kubenswrapper[4714]: I0129 16:29:15.631434 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"1df413aa-ba11-47aa-9b37-989956046c9f","Type":"ContainerStarted","Data":"267d12a21d0575c908ad4f4d9479c0c501da5cdcd406eba9e25bec597bbd362d"} Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.214477 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.401201 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.422941 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-backup-1" podStartSLOduration=6.422909442 podStartE2EDuration="6.422909442s" podCreationTimestamp="2026-01-29 16:29:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:29:15.664635134 +0000 UTC m=+1162.185136284" watchObservedRunningTime="2026-01-29 16:29:19.422909442 +0000 UTC m=+1165.943410562" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.724039 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.727114 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.730811 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.832870 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-config-data\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.833308 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-etc-nvme\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.833353 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpsqq\" (UniqueName: \"kubernetes.io/projected/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-kube-api-access-xpsqq\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.833441 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-etc-machine-id\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.833519 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-var-locks-brick\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.833557 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-lib-modules\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.833593 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-var-locks-cinder\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.833630 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-sys\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.833656 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-config-data-custom\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.833719 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-dev\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.833809 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-etc-iscsi\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.833839 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-run\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.833880 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-scripts\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.834035 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-var-lib-cinder\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.935246 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-lib-modules\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.935324 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-var-locks-cinder\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.935365 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-sys\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.935396 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-config-data-custom\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.935435 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-dev\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.935476 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-dev\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.935395 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-lib-modules\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.935520 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-etc-iscsi\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.935440 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-sys\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.935516 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-var-locks-cinder\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.935549 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-run\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.935595 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-run\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.935608 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-scripts\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.935675 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-etc-iscsi\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.935776 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-var-lib-cinder\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.935686 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-var-lib-cinder\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.935890 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-config-data\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.935977 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-etc-nvme\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.936034 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpsqq\" (UniqueName: \"kubernetes.io/projected/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-kube-api-access-xpsqq\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.936115 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-etc-machine-id\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.936135 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-etc-nvme\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.936252 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-etc-machine-id\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.937008 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-var-locks-brick\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.936828 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-var-locks-brick\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.947349 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-config-data-custom\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.949746 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-scripts\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.950080 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-config-data\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:19 crc kubenswrapper[4714]: I0129 16:29:19.958109 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpsqq\" (UniqueName: \"kubernetes.io/projected/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-kube-api-access-xpsqq\") pod \"cinder-backup-2\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:20 crc kubenswrapper[4714]: I0129 16:29:20.047565 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:20 crc kubenswrapper[4714]: I0129 16:29:20.511482 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Jan 29 16:29:20 crc kubenswrapper[4714]: W0129 16:29:20.517105 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc48b6b3_54d9_4751_a508_b351b1d7fc5d.slice/crio-59d59370e14af6ba69c2be0d33017b635c7ada8221c2372e0298182f5ac46b4b WatchSource:0}: Error finding container 59d59370e14af6ba69c2be0d33017b635c7ada8221c2372e0298182f5ac46b4b: Status 404 returned error can't find the container with id 59d59370e14af6ba69c2be0d33017b635c7ada8221c2372e0298182f5ac46b4b Jan 29 16:29:20 crc kubenswrapper[4714]: I0129 16:29:20.668006 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"cc48b6b3-54d9-4751-a508-b351b1d7fc5d","Type":"ContainerStarted","Data":"59d59370e14af6ba69c2be0d33017b635c7ada8221c2372e0298182f5ac46b4b"} Jan 29 16:29:21 crc kubenswrapper[4714]: E0129 16:29:21.186606 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:29:21 crc kubenswrapper[4714]: I0129 16:29:21.676723 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"cc48b6b3-54d9-4751-a508-b351b1d7fc5d","Type":"ContainerStarted","Data":"3b310f0bce66621749bf8c882b4c5762ba209ef177b7c55ba2a999b2c1e8074a"} Jan 29 16:29:21 crc kubenswrapper[4714]: I0129 16:29:21.676768 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"cc48b6b3-54d9-4751-a508-b351b1d7fc5d","Type":"ContainerStarted","Data":"2a756df7a344d06aa2fb997ad695f4c38702f1186154c912327c3b28f9495911"} Jan 29 16:29:21 crc kubenswrapper[4714]: I0129 16:29:21.699683 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-backup-2" podStartSLOduration=2.699661413 podStartE2EDuration="2.699661413s" podCreationTimestamp="2026-01-29 16:29:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:29:21.696542607 +0000 UTC m=+1168.217043747" watchObservedRunningTime="2026-01-29 16:29:21.699661413 +0000 UTC m=+1168.220162533" Jan 29 16:29:25 crc kubenswrapper[4714]: I0129 16:29:25.048120 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:25 crc kubenswrapper[4714]: I0129 16:29:25.184742 4714 scope.go:117] "RemoveContainer" containerID="f087542636afa3fd11613ddd38d7bd61a4a8cac1161d258a9bca9a6772f2da3c" Jan 29 16:29:25 crc kubenswrapper[4714]: I0129 16:29:25.184779 4714 scope.go:117] "RemoveContainer" containerID="cdf72985bc60f6ce06d587a8a7eb4e0fd4beed9399bf382db190b6ec4763fc7c" Jan 29 16:29:25 crc kubenswrapper[4714]: I0129 16:29:25.299710 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:25 crc kubenswrapper[4714]: I0129 16:29:25.709697 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"2e154d80-4b79-4f74-809e-c1c274ed4063","Type":"ContainerStarted","Data":"06b2f42073c400aedcdcd2bb5ec2c469776d6fe3bfbe8ce136fb5ad93196a673"} Jan 29 16:29:25 crc kubenswrapper[4714]: I0129 16:29:25.710100 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"2e154d80-4b79-4f74-809e-c1c274ed4063","Type":"ContainerStarted","Data":"bbd0c612c943b6ec94ed1610e6f355545ecaf388e165bbb67cb2c301b434e969"} Jan 29 16:29:26 crc kubenswrapper[4714]: I0129 16:29:26.322060 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Jan 29 16:29:26 crc kubenswrapper[4714]: I0129 16:29:26.715812 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-2" podUID="cc48b6b3-54d9-4751-a508-b351b1d7fc5d" containerName="cinder-backup" containerID="cri-o://2a756df7a344d06aa2fb997ad695f4c38702f1186154c912327c3b28f9495911" gracePeriod=30 Jan 29 16:29:26 crc kubenswrapper[4714]: I0129 16:29:26.715902 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-2" podUID="cc48b6b3-54d9-4751-a508-b351b1d7fc5d" containerName="probe" containerID="cri-o://3b310f0bce66621749bf8c882b4c5762ba209ef177b7c55ba2a999b2c1e8074a" gracePeriod=30 Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.714648 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.724458 4714 generic.go:334] "Generic (PLEG): container finished" podID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerID="06b2f42073c400aedcdcd2bb5ec2c469776d6fe3bfbe8ce136fb5ad93196a673" exitCode=1 Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.724502 4714 generic.go:334] "Generic (PLEG): container finished" podID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerID="bbd0c612c943b6ec94ed1610e6f355545ecaf388e165bbb67cb2c301b434e969" exitCode=1 Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.724549 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"2e154d80-4b79-4f74-809e-c1c274ed4063","Type":"ContainerDied","Data":"06b2f42073c400aedcdcd2bb5ec2c469776d6fe3bfbe8ce136fb5ad93196a673"} Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.724582 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"2e154d80-4b79-4f74-809e-c1c274ed4063","Type":"ContainerDied","Data":"bbd0c612c943b6ec94ed1610e6f355545ecaf388e165bbb67cb2c301b434e969"} Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.724602 4714 scope.go:117] "RemoveContainer" containerID="cdf72985bc60f6ce06d587a8a7eb4e0fd4beed9399bf382db190b6ec4763fc7c" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.725251 4714 scope.go:117] "RemoveContainer" containerID="bbd0c612c943b6ec94ed1610e6f355545ecaf388e165bbb67cb2c301b434e969" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.725306 4714 scope.go:117] "RemoveContainer" containerID="06b2f42073c400aedcdcd2bb5ec2c469776d6fe3bfbe8ce136fb5ad93196a673" Jan 29 16:29:27 crc kubenswrapper[4714]: E0129 16:29:27.725656 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.729732 4714 generic.go:334] "Generic (PLEG): container finished" podID="cc48b6b3-54d9-4751-a508-b351b1d7fc5d" containerID="3b310f0bce66621749bf8c882b4c5762ba209ef177b7c55ba2a999b2c1e8074a" exitCode=0 Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.729774 4714 generic.go:334] "Generic (PLEG): container finished" podID="cc48b6b3-54d9-4751-a508-b351b1d7fc5d" containerID="2a756df7a344d06aa2fb997ad695f4c38702f1186154c912327c3b28f9495911" exitCode=0 Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.729816 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"cc48b6b3-54d9-4751-a508-b351b1d7fc5d","Type":"ContainerDied","Data":"3b310f0bce66621749bf8c882b4c5762ba209ef177b7c55ba2a999b2c1e8074a"} Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.729851 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"cc48b6b3-54d9-4751-a508-b351b1d7fc5d","Type":"ContainerDied","Data":"2a756df7a344d06aa2fb997ad695f4c38702f1186154c912327c3b28f9495911"} Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.729864 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"cc48b6b3-54d9-4751-a508-b351b1d7fc5d","Type":"ContainerDied","Data":"59d59370e14af6ba69c2be0d33017b635c7ada8221c2372e0298182f5ac46b4b"} Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.729926 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-2" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.782524 4714 scope.go:117] "RemoveContainer" containerID="f087542636afa3fd11613ddd38d7bd61a4a8cac1161d258a9bca9a6772f2da3c" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.823552 4714 scope.go:117] "RemoveContainer" containerID="3b310f0bce66621749bf8c882b4c5762ba209ef177b7c55ba2a999b2c1e8074a" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.837836 4714 scope.go:117] "RemoveContainer" containerID="2a756df7a344d06aa2fb997ad695f4c38702f1186154c912327c3b28f9495911" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.855747 4714 scope.go:117] "RemoveContainer" containerID="3b310f0bce66621749bf8c882b4c5762ba209ef177b7c55ba2a999b2c1e8074a" Jan 29 16:29:27 crc kubenswrapper[4714]: E0129 16:29:27.856265 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b310f0bce66621749bf8c882b4c5762ba209ef177b7c55ba2a999b2c1e8074a\": container with ID starting with 3b310f0bce66621749bf8c882b4c5762ba209ef177b7c55ba2a999b2c1e8074a not found: ID does not exist" containerID="3b310f0bce66621749bf8c882b4c5762ba209ef177b7c55ba2a999b2c1e8074a" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.856318 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b310f0bce66621749bf8c882b4c5762ba209ef177b7c55ba2a999b2c1e8074a"} err="failed to get container status \"3b310f0bce66621749bf8c882b4c5762ba209ef177b7c55ba2a999b2c1e8074a\": rpc error: code = NotFound desc = could not find container \"3b310f0bce66621749bf8c882b4c5762ba209ef177b7c55ba2a999b2c1e8074a\": container with ID starting with 3b310f0bce66621749bf8c882b4c5762ba209ef177b7c55ba2a999b2c1e8074a not found: ID does not exist" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.856350 4714 scope.go:117] "RemoveContainer" containerID="2a756df7a344d06aa2fb997ad695f4c38702f1186154c912327c3b28f9495911" Jan 29 16:29:27 crc kubenswrapper[4714]: E0129 16:29:27.856693 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a756df7a344d06aa2fb997ad695f4c38702f1186154c912327c3b28f9495911\": container with ID starting with 2a756df7a344d06aa2fb997ad695f4c38702f1186154c912327c3b28f9495911 not found: ID does not exist" containerID="2a756df7a344d06aa2fb997ad695f4c38702f1186154c912327c3b28f9495911" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.856728 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a756df7a344d06aa2fb997ad695f4c38702f1186154c912327c3b28f9495911"} err="failed to get container status \"2a756df7a344d06aa2fb997ad695f4c38702f1186154c912327c3b28f9495911\": rpc error: code = NotFound desc = could not find container \"2a756df7a344d06aa2fb997ad695f4c38702f1186154c912327c3b28f9495911\": container with ID starting with 2a756df7a344d06aa2fb997ad695f4c38702f1186154c912327c3b28f9495911 not found: ID does not exist" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.856752 4714 scope.go:117] "RemoveContainer" containerID="3b310f0bce66621749bf8c882b4c5762ba209ef177b7c55ba2a999b2c1e8074a" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.857099 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b310f0bce66621749bf8c882b4c5762ba209ef177b7c55ba2a999b2c1e8074a"} err="failed to get container status \"3b310f0bce66621749bf8c882b4c5762ba209ef177b7c55ba2a999b2c1e8074a\": rpc error: code = NotFound desc = could not find container \"3b310f0bce66621749bf8c882b4c5762ba209ef177b7c55ba2a999b2c1e8074a\": container with ID starting with 3b310f0bce66621749bf8c882b4c5762ba209ef177b7c55ba2a999b2c1e8074a not found: ID does not exist" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.857120 4714 scope.go:117] "RemoveContainer" containerID="2a756df7a344d06aa2fb997ad695f4c38702f1186154c912327c3b28f9495911" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.857352 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a756df7a344d06aa2fb997ad695f4c38702f1186154c912327c3b28f9495911"} err="failed to get container status \"2a756df7a344d06aa2fb997ad695f4c38702f1186154c912327c3b28f9495911\": rpc error: code = NotFound desc = could not find container \"2a756df7a344d06aa2fb997ad695f4c38702f1186154c912327c3b28f9495911\": container with ID starting with 2a756df7a344d06aa2fb997ad695f4c38702f1186154c912327c3b28f9495911 not found: ID does not exist" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.869183 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-etc-iscsi\") pod \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.869240 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-config-data-custom\") pod \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.869260 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-etc-nvme\") pod \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.869285 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-sys\") pod \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.869322 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpsqq\" (UniqueName: \"kubernetes.io/projected/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-kube-api-access-xpsqq\") pod \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.869339 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-lib-modules\") pod \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.869340 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "cc48b6b3-54d9-4751-a508-b351b1d7fc5d" (UID: "cc48b6b3-54d9-4751-a508-b351b1d7fc5d"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.869370 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-run\") pod \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.869390 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-var-locks-cinder\") pod \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.869407 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-sys" (OuterVolumeSpecName: "sys") pod "cc48b6b3-54d9-4751-a508-b351b1d7fc5d" (UID: "cc48b6b3-54d9-4751-a508-b351b1d7fc5d"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.869437 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-var-locks-brick\") pod \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.869462 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-config-data\") pod \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.869483 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-dev\") pod \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.869510 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-scripts\") pod \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.869557 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-etc-machine-id\") pod \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.869578 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-var-lib-cinder\") pod \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\" (UID: \"cc48b6b3-54d9-4751-a508-b351b1d7fc5d\") " Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.869844 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "cc48b6b3-54d9-4751-a508-b351b1d7fc5d" (UID: "cc48b6b3-54d9-4751-a508-b351b1d7fc5d"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.870156 4714 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-etc-iscsi\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.870174 4714 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-etc-nvme\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.870188 4714 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-sys\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.870301 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-dev" (OuterVolumeSpecName: "dev") pod "cc48b6b3-54d9-4751-a508-b351b1d7fc5d" (UID: "cc48b6b3-54d9-4751-a508-b351b1d7fc5d"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.870328 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "cc48b6b3-54d9-4751-a508-b351b1d7fc5d" (UID: "cc48b6b3-54d9-4751-a508-b351b1d7fc5d"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.870356 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "cc48b6b3-54d9-4751-a508-b351b1d7fc5d" (UID: "cc48b6b3-54d9-4751-a508-b351b1d7fc5d"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.870387 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "cc48b6b3-54d9-4751-a508-b351b1d7fc5d" (UID: "cc48b6b3-54d9-4751-a508-b351b1d7fc5d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.870414 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "cc48b6b3-54d9-4751-a508-b351b1d7fc5d" (UID: "cc48b6b3-54d9-4751-a508-b351b1d7fc5d"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.870415 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-run" (OuterVolumeSpecName: "run") pod "cc48b6b3-54d9-4751-a508-b351b1d7fc5d" (UID: "cc48b6b3-54d9-4751-a508-b351b1d7fc5d"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.870435 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "cc48b6b3-54d9-4751-a508-b351b1d7fc5d" (UID: "cc48b6b3-54d9-4751-a508-b351b1d7fc5d"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.874454 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-scripts" (OuterVolumeSpecName: "scripts") pod "cc48b6b3-54d9-4751-a508-b351b1d7fc5d" (UID: "cc48b6b3-54d9-4751-a508-b351b1d7fc5d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.874468 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-kube-api-access-xpsqq" (OuterVolumeSpecName: "kube-api-access-xpsqq") pod "cc48b6b3-54d9-4751-a508-b351b1d7fc5d" (UID: "cc48b6b3-54d9-4751-a508-b351b1d7fc5d"). InnerVolumeSpecName "kube-api-access-xpsqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.874629 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cc48b6b3-54d9-4751-a508-b351b1d7fc5d" (UID: "cc48b6b3-54d9-4751-a508-b351b1d7fc5d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.936522 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-config-data" (OuterVolumeSpecName: "config-data") pod "cc48b6b3-54d9-4751-a508-b351b1d7fc5d" (UID: "cc48b6b3-54d9-4751-a508-b351b1d7fc5d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.971508 4714 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.971549 4714 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.971564 4714 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.971580 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.971593 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpsqq\" (UniqueName: \"kubernetes.io/projected/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-kube-api-access-xpsqq\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.971607 4714 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-lib-modules\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.971619 4714 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-run\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.971633 4714 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.971645 4714 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-var-locks-brick\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.971657 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:27 crc kubenswrapper[4714]: I0129 16:29:27.971668 4714 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cc48b6b3-54d9-4751-a508-b351b1d7fc5d-dev\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:28 crc kubenswrapper[4714]: I0129 16:29:28.067138 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Jan 29 16:29:28 crc kubenswrapper[4714]: I0129 16:29:28.082511 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Jan 29 16:29:28 crc kubenswrapper[4714]: I0129 16:29:28.088831 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Jan 29 16:29:28 crc kubenswrapper[4714]: I0129 16:29:28.089254 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-1" podUID="1df413aa-ba11-47aa-9b37-989956046c9f" containerName="probe" containerID="cri-o://216659e8b1aeed2e1c0a3e238db1d9a8366df5059f31bffd8d1c1116d3eda51f" gracePeriod=30 Jan 29 16:29:28 crc kubenswrapper[4714]: I0129 16:29:28.089237 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-1" podUID="1df413aa-ba11-47aa-9b37-989956046c9f" containerName="cinder-backup" containerID="cri-o://11899e2f2c2124b1f13dd44e2372c23241e6ddb29215a4c11ae4fe8ed891c9bb" gracePeriod=30 Jan 29 16:29:28 crc kubenswrapper[4714]: I0129 16:29:28.194486 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc48b6b3-54d9-4751-a508-b351b1d7fc5d" path="/var/lib/kubelet/pods/cc48b6b3-54d9-4751-a508-b351b1d7fc5d/volumes" Jan 29 16:29:28 crc kubenswrapper[4714]: I0129 16:29:28.745308 4714 generic.go:334] "Generic (PLEG): container finished" podID="1df413aa-ba11-47aa-9b37-989956046c9f" containerID="216659e8b1aeed2e1c0a3e238db1d9a8366df5059f31bffd8d1c1116d3eda51f" exitCode=0 Jan 29 16:29:28 crc kubenswrapper[4714]: I0129 16:29:28.745402 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"1df413aa-ba11-47aa-9b37-989956046c9f","Type":"ContainerDied","Data":"216659e8b1aeed2e1c0a3e238db1d9a8366df5059f31bffd8d1c1116d3eda51f"} Jan 29 16:29:29 crc kubenswrapper[4714]: I0129 16:29:29.887304 4714 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:29:29 crc kubenswrapper[4714]: I0129 16:29:29.887562 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:29:29 crc kubenswrapper[4714]: I0129 16:29:29.887572 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:29:29 crc kubenswrapper[4714]: I0129 16:29:29.888126 4714 scope.go:117] "RemoveContainer" containerID="bbd0c612c943b6ec94ed1610e6f355545ecaf388e165bbb67cb2c301b434e969" Jan 29 16:29:29 crc kubenswrapper[4714]: I0129 16:29:29.888137 4714 scope.go:117] "RemoveContainer" containerID="06b2f42073c400aedcdcd2bb5ec2c469776d6fe3bfbe8ce136fb5ad93196a673" Jan 29 16:29:29 crc kubenswrapper[4714]: E0129 16:29:29.888406 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" Jan 29 16:29:31 crc kubenswrapper[4714]: I0129 16:29:31.776526 4714 generic.go:334] "Generic (PLEG): container finished" podID="1df413aa-ba11-47aa-9b37-989956046c9f" containerID="11899e2f2c2124b1f13dd44e2372c23241e6ddb29215a4c11ae4fe8ed891c9bb" exitCode=0 Jan 29 16:29:31 crc kubenswrapper[4714]: I0129 16:29:31.776666 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"1df413aa-ba11-47aa-9b37-989956046c9f","Type":"ContainerDied","Data":"11899e2f2c2124b1f13dd44e2372c23241e6ddb29215a4c11ae4fe8ed891c9bb"} Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.243008 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.350234 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1df413aa-ba11-47aa-9b37-989956046c9f-scripts\") pod \"1df413aa-ba11-47aa-9b37-989956046c9f\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.350296 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-run\") pod \"1df413aa-ba11-47aa-9b37-989956046c9f\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.350325 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-lib-modules\") pod \"1df413aa-ba11-47aa-9b37-989956046c9f\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.350376 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmcjf\" (UniqueName: \"kubernetes.io/projected/1df413aa-ba11-47aa-9b37-989956046c9f-kube-api-access-kmcjf\") pod \"1df413aa-ba11-47aa-9b37-989956046c9f\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.350393 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-etc-machine-id\") pod \"1df413aa-ba11-47aa-9b37-989956046c9f\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.350409 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1df413aa-ba11-47aa-9b37-989956046c9f-config-data\") pod \"1df413aa-ba11-47aa-9b37-989956046c9f\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.350426 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-etc-iscsi\") pod \"1df413aa-ba11-47aa-9b37-989956046c9f\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.350457 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-var-locks-cinder\") pod \"1df413aa-ba11-47aa-9b37-989956046c9f\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.350476 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-dev\") pod \"1df413aa-ba11-47aa-9b37-989956046c9f\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.350500 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1df413aa-ba11-47aa-9b37-989956046c9f-config-data-custom\") pod \"1df413aa-ba11-47aa-9b37-989956046c9f\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.350559 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-sys\") pod \"1df413aa-ba11-47aa-9b37-989956046c9f\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.350575 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-etc-nvme\") pod \"1df413aa-ba11-47aa-9b37-989956046c9f\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.350589 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-var-locks-brick\") pod \"1df413aa-ba11-47aa-9b37-989956046c9f\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.350615 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-var-lib-cinder\") pod \"1df413aa-ba11-47aa-9b37-989956046c9f\" (UID: \"1df413aa-ba11-47aa-9b37-989956046c9f\") " Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.351491 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "1df413aa-ba11-47aa-9b37-989956046c9f" (UID: "1df413aa-ba11-47aa-9b37-989956046c9f"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.351582 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "1df413aa-ba11-47aa-9b37-989956046c9f" (UID: "1df413aa-ba11-47aa-9b37-989956046c9f"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.351589 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "1df413aa-ba11-47aa-9b37-989956046c9f" (UID: "1df413aa-ba11-47aa-9b37-989956046c9f"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.351623 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-dev" (OuterVolumeSpecName: "dev") pod "1df413aa-ba11-47aa-9b37-989956046c9f" (UID: "1df413aa-ba11-47aa-9b37-989956046c9f"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.351643 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-sys" (OuterVolumeSpecName: "sys") pod "1df413aa-ba11-47aa-9b37-989956046c9f" (UID: "1df413aa-ba11-47aa-9b37-989956046c9f"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.351640 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "1df413aa-ba11-47aa-9b37-989956046c9f" (UID: "1df413aa-ba11-47aa-9b37-989956046c9f"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.351669 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-run" (OuterVolumeSpecName: "run") pod "1df413aa-ba11-47aa-9b37-989956046c9f" (UID: "1df413aa-ba11-47aa-9b37-989956046c9f"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.351670 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "1df413aa-ba11-47aa-9b37-989956046c9f" (UID: "1df413aa-ba11-47aa-9b37-989956046c9f"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.351689 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1df413aa-ba11-47aa-9b37-989956046c9f" (UID: "1df413aa-ba11-47aa-9b37-989956046c9f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.352332 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "1df413aa-ba11-47aa-9b37-989956046c9f" (UID: "1df413aa-ba11-47aa-9b37-989956046c9f"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.370371 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1df413aa-ba11-47aa-9b37-989956046c9f-scripts" (OuterVolumeSpecName: "scripts") pod "1df413aa-ba11-47aa-9b37-989956046c9f" (UID: "1df413aa-ba11-47aa-9b37-989956046c9f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.370389 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1df413aa-ba11-47aa-9b37-989956046c9f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1df413aa-ba11-47aa-9b37-989956046c9f" (UID: "1df413aa-ba11-47aa-9b37-989956046c9f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.370425 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1df413aa-ba11-47aa-9b37-989956046c9f-kube-api-access-kmcjf" (OuterVolumeSpecName: "kube-api-access-kmcjf") pod "1df413aa-ba11-47aa-9b37-989956046c9f" (UID: "1df413aa-ba11-47aa-9b37-989956046c9f"). InnerVolumeSpecName "kube-api-access-kmcjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.438943 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1df413aa-ba11-47aa-9b37-989956046c9f-config-data" (OuterVolumeSpecName: "config-data") pod "1df413aa-ba11-47aa-9b37-989956046c9f" (UID: "1df413aa-ba11-47aa-9b37-989956046c9f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.453142 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmcjf\" (UniqueName: \"kubernetes.io/projected/1df413aa-ba11-47aa-9b37-989956046c9f-kube-api-access-kmcjf\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.453188 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1df413aa-ba11-47aa-9b37-989956046c9f-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.453204 4714 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.453216 4714 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-etc-iscsi\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.453229 4714 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.453240 4714 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-dev\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.453252 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1df413aa-ba11-47aa-9b37-989956046c9f-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.453263 4714 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-sys\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.453275 4714 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-etc-nvme\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.453287 4714 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-var-locks-brick\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.453299 4714 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.453311 4714 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1df413aa-ba11-47aa-9b37-989956046c9f-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.453322 4714 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-run\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.453333 4714 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1df413aa-ba11-47aa-9b37-989956046c9f-lib-modules\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.788003 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"1df413aa-ba11-47aa-9b37-989956046c9f","Type":"ContainerDied","Data":"267d12a21d0575c908ad4f4d9479c0c501da5cdcd406eba9e25bec597bbd362d"} Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.788379 4714 scope.go:117] "RemoveContainer" containerID="216659e8b1aeed2e1c0a3e238db1d9a8366df5059f31bffd8d1c1116d3eda51f" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.788110 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-1" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.817045 4714 scope.go:117] "RemoveContainer" containerID="11899e2f2c2124b1f13dd44e2372c23241e6ddb29215a4c11ae4fe8ed891c9bb" Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.848876 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Jan 29 16:29:32 crc kubenswrapper[4714]: I0129 16:29:32.854074 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Jan 29 16:29:33 crc kubenswrapper[4714]: I0129 16:29:33.616022 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Jan 29 16:29:33 crc kubenswrapper[4714]: I0129 16:29:33.616654 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3" containerName="cinder-api-log" containerID="cri-o://79461c5a69e4029a68e7e8faa95d70a67ee30b1d8f5d9b8793dd7214e3da2c42" gracePeriod=30 Jan 29 16:29:33 crc kubenswrapper[4714]: I0129 16:29:33.616786 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3" containerName="cinder-api" containerID="cri-o://e1b4f5a218866bd462e639ef9e2c44453f1907feb94c13e36f29e0c48f866074" gracePeriod=30 Jan 29 16:29:33 crc kubenswrapper[4714]: I0129 16:29:33.794699 4714 generic.go:334] "Generic (PLEG): container finished" podID="9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3" containerID="79461c5a69e4029a68e7e8faa95d70a67ee30b1d8f5d9b8793dd7214e3da2c42" exitCode=143 Jan 29 16:29:33 crc kubenswrapper[4714]: I0129 16:29:33.794761 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3","Type":"ContainerDied","Data":"79461c5a69e4029a68e7e8faa95d70a67ee30b1d8f5d9b8793dd7214e3da2c42"} Jan 29 16:29:34 crc kubenswrapper[4714]: I0129 16:29:34.198459 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1df413aa-ba11-47aa-9b37-989956046c9f" path="/var/lib/kubelet/pods/1df413aa-ba11-47aa-9b37-989956046c9f/volumes" Jan 29 16:29:35 crc kubenswrapper[4714]: E0129 16:29:35.188150 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:29:36 crc kubenswrapper[4714]: I0129 16:29:36.766656 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/cinder-api-0" podUID="9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.93:8776/healthcheck\": read tcp 10.217.0.2:49470->10.217.0.93:8776: read: connection reset by peer" Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.745804 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.828616 4714 generic.go:334] "Generic (PLEG): container finished" podID="9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3" containerID="e1b4f5a218866bd462e639ef9e2c44453f1907feb94c13e36f29e0c48f866074" exitCode=0 Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.828662 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3","Type":"ContainerDied","Data":"e1b4f5a218866bd462e639ef9e2c44453f1907feb94c13e36f29e0c48f866074"} Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.828690 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3","Type":"ContainerDied","Data":"4d4f411f7f92df4f5d547c75bfc715baa406061177023b383b53537927256cc7"} Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.828706 4714 scope.go:117] "RemoveContainer" containerID="e1b4f5a218866bd462e639ef9e2c44453f1907feb94c13e36f29e0c48f866074" Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.828835 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.849762 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-config-data-custom\") pod \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.849870 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-796r4\" (UniqueName: \"kubernetes.io/projected/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-kube-api-access-796r4\") pod \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.849961 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-scripts\") pod \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.849988 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-logs\") pod \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.850013 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-config-data\") pod \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.850038 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-etc-machine-id\") pod \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\" (UID: \"9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3\") " Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.850320 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3" (UID: "9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.851213 4714 scope.go:117] "RemoveContainer" containerID="79461c5a69e4029a68e7e8faa95d70a67ee30b1d8f5d9b8793dd7214e3da2c42" Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.851615 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-logs" (OuterVolumeSpecName: "logs") pod "9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3" (UID: "9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.858370 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-scripts" (OuterVolumeSpecName: "scripts") pod "9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3" (UID: "9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.861334 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-kube-api-access-796r4" (OuterVolumeSpecName: "kube-api-access-796r4") pod "9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3" (UID: "9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3"). InnerVolumeSpecName "kube-api-access-796r4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.867201 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3" (UID: "9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.891079 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-config-data" (OuterVolumeSpecName: "config-data") pod "9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3" (UID: "9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.944133 4714 scope.go:117] "RemoveContainer" containerID="e1b4f5a218866bd462e639ef9e2c44453f1907feb94c13e36f29e0c48f866074" Jan 29 16:29:37 crc kubenswrapper[4714]: E0129 16:29:37.944618 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1b4f5a218866bd462e639ef9e2c44453f1907feb94c13e36f29e0c48f866074\": container with ID starting with e1b4f5a218866bd462e639ef9e2c44453f1907feb94c13e36f29e0c48f866074 not found: ID does not exist" containerID="e1b4f5a218866bd462e639ef9e2c44453f1907feb94c13e36f29e0c48f866074" Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.944662 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1b4f5a218866bd462e639ef9e2c44453f1907feb94c13e36f29e0c48f866074"} err="failed to get container status \"e1b4f5a218866bd462e639ef9e2c44453f1907feb94c13e36f29e0c48f866074\": rpc error: code = NotFound desc = could not find container \"e1b4f5a218866bd462e639ef9e2c44453f1907feb94c13e36f29e0c48f866074\": container with ID starting with e1b4f5a218866bd462e639ef9e2c44453f1907feb94c13e36f29e0c48f866074 not found: ID does not exist" Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.944690 4714 scope.go:117] "RemoveContainer" containerID="79461c5a69e4029a68e7e8faa95d70a67ee30b1d8f5d9b8793dd7214e3da2c42" Jan 29 16:29:37 crc kubenswrapper[4714]: E0129 16:29:37.944959 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79461c5a69e4029a68e7e8faa95d70a67ee30b1d8f5d9b8793dd7214e3da2c42\": container with ID starting with 79461c5a69e4029a68e7e8faa95d70a67ee30b1d8f5d9b8793dd7214e3da2c42 not found: ID does not exist" containerID="79461c5a69e4029a68e7e8faa95d70a67ee30b1d8f5d9b8793dd7214e3da2c42" Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.944988 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79461c5a69e4029a68e7e8faa95d70a67ee30b1d8f5d9b8793dd7214e3da2c42"} err="failed to get container status \"79461c5a69e4029a68e7e8faa95d70a67ee30b1d8f5d9b8793dd7214e3da2c42\": rpc error: code = NotFound desc = could not find container \"79461c5a69e4029a68e7e8faa95d70a67ee30b1d8f5d9b8793dd7214e3da2c42\": container with ID starting with 79461c5a69e4029a68e7e8faa95d70a67ee30b1d8f5d9b8793dd7214e3da2c42 not found: ID does not exist" Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.952100 4714 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-logs\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.952123 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.952132 4714 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.952140 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.952149 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-796r4\" (UniqueName: \"kubernetes.io/projected/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-kube-api-access-796r4\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:37 crc kubenswrapper[4714]: I0129 16:29:37.952157 4714 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.166049 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.174067 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.198684 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3" path="/var/lib/kubelet/pods/9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3/volumes" Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.945624 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Jan 29 16:29:38 crc kubenswrapper[4714]: E0129 16:29:38.946621 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3" containerName="cinder-api-log" Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.946652 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3" containerName="cinder-api-log" Jan 29 16:29:38 crc kubenswrapper[4714]: E0129 16:29:38.946683 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3" containerName="cinder-api" Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.946700 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3" containerName="cinder-api" Jan 29 16:29:38 crc kubenswrapper[4714]: E0129 16:29:38.946718 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc48b6b3-54d9-4751-a508-b351b1d7fc5d" containerName="cinder-backup" Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.946733 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc48b6b3-54d9-4751-a508-b351b1d7fc5d" containerName="cinder-backup" Jan 29 16:29:38 crc kubenswrapper[4714]: E0129 16:29:38.946774 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df413aa-ba11-47aa-9b37-989956046c9f" containerName="cinder-backup" Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.946789 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df413aa-ba11-47aa-9b37-989956046c9f" containerName="cinder-backup" Jan 29 16:29:38 crc kubenswrapper[4714]: E0129 16:29:38.946822 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc48b6b3-54d9-4751-a508-b351b1d7fc5d" containerName="probe" Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.946837 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc48b6b3-54d9-4751-a508-b351b1d7fc5d" containerName="probe" Jan 29 16:29:38 crc kubenswrapper[4714]: E0129 16:29:38.946869 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df413aa-ba11-47aa-9b37-989956046c9f" containerName="probe" Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.946886 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df413aa-ba11-47aa-9b37-989956046c9f" containerName="probe" Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.947234 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3" containerName="cinder-api" Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.947257 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="1df413aa-ba11-47aa-9b37-989956046c9f" containerName="cinder-backup" Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.947274 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc48b6b3-54d9-4751-a508-b351b1d7fc5d" containerName="probe" Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.947302 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b21fd86-5b9b-4b13-82aa-eb3d7f1fafb3" containerName="cinder-api-log" Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.947329 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="1df413aa-ba11-47aa-9b37-989956046c9f" containerName="probe" Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.947356 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc48b6b3-54d9-4751-a508-b351b1d7fc5d" containerName="cinder-backup" Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.948913 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.952259 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.953374 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.962493 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.963813 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:38 crc kubenswrapper[4714]: I0129 16:29:38.968226 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-api-config-data" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.067637 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.072164 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5a3f592-54c6-44f1-9f09-f366502287a6-config-data-custom\") pod \"cinder-api-0\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.072212 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/869a701a-040e-44ea-98cc-53eb5f33c933-scripts\") pod \"cinder-api-2\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.072235 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk7jf\" (UniqueName: \"kubernetes.io/projected/84becd41-a7aa-4b36-a6a7-8516c2e64909-kube-api-access-xk7jf\") pod \"cinder-api-1\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.072253 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p55ww\" (UniqueName: \"kubernetes.io/projected/c5a3f592-54c6-44f1-9f09-f366502287a6-kube-api-access-p55ww\") pod \"cinder-api-0\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.072269 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/869a701a-040e-44ea-98cc-53eb5f33c933-config-data\") pod \"cinder-api-2\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.072286 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84becd41-a7aa-4b36-a6a7-8516c2e64909-config-data\") pod \"cinder-api-1\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.072306 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5a3f592-54c6-44f1-9f09-f366502287a6-scripts\") pod \"cinder-api-0\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.072327 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/869a701a-040e-44ea-98cc-53eb5f33c933-config-data-custom\") pod \"cinder-api-2\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.072353 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/84becd41-a7aa-4b36-a6a7-8516c2e64909-etc-machine-id\") pod \"cinder-api-1\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.072371 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c5a3f592-54c6-44f1-9f09-f366502287a6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.072385 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84becd41-a7aa-4b36-a6a7-8516c2e64909-scripts\") pod \"cinder-api-1\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.072399 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/869a701a-040e-44ea-98cc-53eb5f33c933-etc-machine-id\") pod \"cinder-api-2\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.072414 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84becd41-a7aa-4b36-a6a7-8516c2e64909-config-data-custom\") pod \"cinder-api-1\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.072432 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84becd41-a7aa-4b36-a6a7-8516c2e64909-logs\") pod \"cinder-api-1\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.072446 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5a3f592-54c6-44f1-9f09-f366502287a6-config-data\") pod \"cinder-api-0\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.072462 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5a3f592-54c6-44f1-9f09-f366502287a6-logs\") pod \"cinder-api-0\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.072476 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/869a701a-040e-44ea-98cc-53eb5f33c933-logs\") pod \"cinder-api-2\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.072515 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zvnq\" (UniqueName: \"kubernetes.io/projected/869a701a-040e-44ea-98cc-53eb5f33c933-kube-api-access-4zvnq\") pod \"cinder-api-2\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.079322 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.087008 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.173708 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/869a701a-040e-44ea-98cc-53eb5f33c933-etc-machine-id\") pod \"cinder-api-2\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.173763 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84becd41-a7aa-4b36-a6a7-8516c2e64909-config-data-custom\") pod \"cinder-api-1\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.173781 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84becd41-a7aa-4b36-a6a7-8516c2e64909-logs\") pod \"cinder-api-1\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.173803 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5a3f592-54c6-44f1-9f09-f366502287a6-config-data\") pod \"cinder-api-0\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.173822 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5a3f592-54c6-44f1-9f09-f366502287a6-logs\") pod \"cinder-api-0\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.173837 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/869a701a-040e-44ea-98cc-53eb5f33c933-logs\") pod \"cinder-api-2\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.173877 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zvnq\" (UniqueName: \"kubernetes.io/projected/869a701a-040e-44ea-98cc-53eb5f33c933-kube-api-access-4zvnq\") pod \"cinder-api-2\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.173917 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5a3f592-54c6-44f1-9f09-f366502287a6-config-data-custom\") pod \"cinder-api-0\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.173958 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/869a701a-040e-44ea-98cc-53eb5f33c933-scripts\") pod \"cinder-api-2\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.173984 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk7jf\" (UniqueName: \"kubernetes.io/projected/84becd41-a7aa-4b36-a6a7-8516c2e64909-kube-api-access-xk7jf\") pod \"cinder-api-1\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.174022 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p55ww\" (UniqueName: \"kubernetes.io/projected/c5a3f592-54c6-44f1-9f09-f366502287a6-kube-api-access-p55ww\") pod \"cinder-api-0\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.174042 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/869a701a-040e-44ea-98cc-53eb5f33c933-config-data\") pod \"cinder-api-2\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.174062 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84becd41-a7aa-4b36-a6a7-8516c2e64909-config-data\") pod \"cinder-api-1\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.174078 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5a3f592-54c6-44f1-9f09-f366502287a6-scripts\") pod \"cinder-api-0\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.174100 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/869a701a-040e-44ea-98cc-53eb5f33c933-config-data-custom\") pod \"cinder-api-2\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.174134 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/84becd41-a7aa-4b36-a6a7-8516c2e64909-etc-machine-id\") pod \"cinder-api-1\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.174161 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c5a3f592-54c6-44f1-9f09-f366502287a6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.174179 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84becd41-a7aa-4b36-a6a7-8516c2e64909-scripts\") pod \"cinder-api-1\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.174508 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/869a701a-040e-44ea-98cc-53eb5f33c933-etc-machine-id\") pod \"cinder-api-2\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.175450 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84becd41-a7aa-4b36-a6a7-8516c2e64909-logs\") pod \"cinder-api-1\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.176511 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5a3f592-54c6-44f1-9f09-f366502287a6-logs\") pod \"cinder-api-0\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.177312 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/869a701a-040e-44ea-98cc-53eb5f33c933-logs\") pod \"cinder-api-2\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.180303 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/84becd41-a7aa-4b36-a6a7-8516c2e64909-etc-machine-id\") pod \"cinder-api-1\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.181385 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/869a701a-040e-44ea-98cc-53eb5f33c933-scripts\") pod \"cinder-api-2\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.181463 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c5a3f592-54c6-44f1-9f09-f366502287a6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.187362 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5a3f592-54c6-44f1-9f09-f366502287a6-config-data-custom\") pod \"cinder-api-0\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.188406 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84becd41-a7aa-4b36-a6a7-8516c2e64909-scripts\") pod \"cinder-api-1\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.188374 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84becd41-a7aa-4b36-a6a7-8516c2e64909-config-data\") pod \"cinder-api-1\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.188562 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/869a701a-040e-44ea-98cc-53eb5f33c933-config-data-custom\") pod \"cinder-api-2\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.188680 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84becd41-a7aa-4b36-a6a7-8516c2e64909-config-data-custom\") pod \"cinder-api-1\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.189351 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5a3f592-54c6-44f1-9f09-f366502287a6-config-data\") pod \"cinder-api-0\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.190513 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/869a701a-040e-44ea-98cc-53eb5f33c933-config-data\") pod \"cinder-api-2\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.194068 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5a3f592-54c6-44f1-9f09-f366502287a6-scripts\") pod \"cinder-api-0\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.194090 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zvnq\" (UniqueName: \"kubernetes.io/projected/869a701a-040e-44ea-98cc-53eb5f33c933-kube-api-access-4zvnq\") pod \"cinder-api-2\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.202049 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk7jf\" (UniqueName: \"kubernetes.io/projected/84becd41-a7aa-4b36-a6a7-8516c2e64909-kube-api-access-xk7jf\") pod \"cinder-api-1\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.202650 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p55ww\" (UniqueName: \"kubernetes.io/projected/c5a3f592-54c6-44f1-9f09-f366502287a6-kube-api-access-p55ww\") pod \"cinder-api-0\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.276376 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.287259 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.296000 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.733708 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Jan 29 16:29:39 crc kubenswrapper[4714]: W0129 16:29:39.741675 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5a3f592_54c6_44f1_9f09_f366502287a6.slice/crio-fcf337a11251f351330506928dad191d61b9373e215ea14ec246e54e5e9a3034 WatchSource:0}: Error finding container fcf337a11251f351330506928dad191d61b9373e215ea14ec246e54e5e9a3034: Status 404 returned error can't find the container with id fcf337a11251f351330506928dad191d61b9373e215ea14ec246e54e5e9a3034 Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.796839 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.812336 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Jan 29 16:29:39 crc kubenswrapper[4714]: W0129 16:29:39.820782 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84becd41_a7aa_4b36_a6a7_8516c2e64909.slice/crio-7f9cbec8d1d3b10670a11a02aa0992c33a6443ac29d99038a22ff077c3971712 WatchSource:0}: Error finding container 7f9cbec8d1d3b10670a11a02aa0992c33a6443ac29d99038a22ff077c3971712: Status 404 returned error can't find the container with id 7f9cbec8d1d3b10670a11a02aa0992c33a6443ac29d99038a22ff077c3971712 Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.877134 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"869a701a-040e-44ea-98cc-53eb5f33c933","Type":"ContainerStarted","Data":"2504822ce0a99b2d0cc7db311f85a30c6e6f161cbc200d841e855f4f3eb1ff5b"} Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.879173 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"84becd41-a7aa-4b36-a6a7-8516c2e64909","Type":"ContainerStarted","Data":"7f9cbec8d1d3b10670a11a02aa0992c33a6443ac29d99038a22ff077c3971712"} Jan 29 16:29:39 crc kubenswrapper[4714]: I0129 16:29:39.881843 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"c5a3f592-54c6-44f1-9f09-f366502287a6","Type":"ContainerStarted","Data":"fcf337a11251f351330506928dad191d61b9373e215ea14ec246e54e5e9a3034"} Jan 29 16:29:40 crc kubenswrapper[4714]: I0129 16:29:40.891643 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"869a701a-040e-44ea-98cc-53eb5f33c933","Type":"ContainerStarted","Data":"fe14e97e807f7bfc78569e0b132f84ed0512e6e44920a45bdbbf02eab4f1fb0d"} Jan 29 16:29:40 crc kubenswrapper[4714]: I0129 16:29:40.894290 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"84becd41-a7aa-4b36-a6a7-8516c2e64909","Type":"ContainerStarted","Data":"df6868e9d7ac6b0ebfef8c6a4c95fca607fa88217512148f0e9154c25064c11e"} Jan 29 16:29:40 crc kubenswrapper[4714]: I0129 16:29:40.898156 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"c5a3f592-54c6-44f1-9f09-f366502287a6","Type":"ContainerStarted","Data":"32d98c19e9cd977d60b8a6256b941e6d21b55abe95ffb10609d846a79c267c86"} Jan 29 16:29:41 crc kubenswrapper[4714]: I0129 16:29:41.184891 4714 scope.go:117] "RemoveContainer" containerID="bbd0c612c943b6ec94ed1610e6f355545ecaf388e165bbb67cb2c301b434e969" Jan 29 16:29:41 crc kubenswrapper[4714]: I0129 16:29:41.185488 4714 scope.go:117] "RemoveContainer" containerID="06b2f42073c400aedcdcd2bb5ec2c469776d6fe3bfbe8ce136fb5ad93196a673" Jan 29 16:29:41 crc kubenswrapper[4714]: E0129 16:29:41.186055 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" Jan 29 16:29:41 crc kubenswrapper[4714]: I0129 16:29:41.914189 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"869a701a-040e-44ea-98cc-53eb5f33c933","Type":"ContainerStarted","Data":"ee0d2c74772fa7181a1d5113e251172e052dd7202e52d6c038b2533ca027dac1"} Jan 29 16:29:41 crc kubenswrapper[4714]: I0129 16:29:41.914360 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:41 crc kubenswrapper[4714]: I0129 16:29:41.918021 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"84becd41-a7aa-4b36-a6a7-8516c2e64909","Type":"ContainerStarted","Data":"59d26536c5d18af3cf0b3379093a52042043c35adcb85fba86b4fe5fa5b56cb8"} Jan 29 16:29:41 crc kubenswrapper[4714]: I0129 16:29:41.918232 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:41 crc kubenswrapper[4714]: I0129 16:29:41.922533 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"c5a3f592-54c6-44f1-9f09-f366502287a6","Type":"ContainerStarted","Data":"922b00d2d16cf09fabe14b15b7a7648c66244d5a615d55c77b2cc333c2095cf3"} Jan 29 16:29:41 crc kubenswrapper[4714]: I0129 16:29:41.923148 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:41 crc kubenswrapper[4714]: I0129 16:29:41.943612 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-2" podStartSLOduration=3.943591689 podStartE2EDuration="3.943591689s" podCreationTimestamp="2026-01-29 16:29:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:29:41.938226617 +0000 UTC m=+1188.458727777" watchObservedRunningTime="2026-01-29 16:29:41.943591689 +0000 UTC m=+1188.464092819" Jan 29 16:29:41 crc kubenswrapper[4714]: I0129 16:29:41.974980 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-0" podStartSLOduration=3.974908069 podStartE2EDuration="3.974908069s" podCreationTimestamp="2026-01-29 16:29:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:29:41.967041803 +0000 UTC m=+1188.487542943" watchObservedRunningTime="2026-01-29 16:29:41.974908069 +0000 UTC m=+1188.495409239" Jan 29 16:29:42 crc kubenswrapper[4714]: I0129 16:29:41.999615 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-1" podStartSLOduration=3.999586308 podStartE2EDuration="3.999586308s" podCreationTimestamp="2026-01-29 16:29:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:29:41.991071709 +0000 UTC m=+1188.511572889" watchObservedRunningTime="2026-01-29 16:29:41.999586308 +0000 UTC m=+1188.520087438" Jan 29 16:29:48 crc kubenswrapper[4714]: E0129 16:29:48.196785 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:29:51 crc kubenswrapper[4714]: I0129 16:29:51.080495 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:29:51 crc kubenswrapper[4714]: I0129 16:29:51.092381 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:51 crc kubenswrapper[4714]: I0129 16:29:51.137606 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:52 crc kubenswrapper[4714]: I0129 16:29:52.185063 4714 scope.go:117] "RemoveContainer" containerID="bbd0c612c943b6ec94ed1610e6f355545ecaf388e165bbb67cb2c301b434e969" Jan 29 16:29:52 crc kubenswrapper[4714]: I0129 16:29:52.185119 4714 scope.go:117] "RemoveContainer" containerID="06b2f42073c400aedcdcd2bb5ec2c469776d6fe3bfbe8ce136fb5ad93196a673" Jan 29 16:29:52 crc kubenswrapper[4714]: E0129 16:29:52.185521 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(2e154d80-4b79-4f74-809e-c1c274ed4063)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" Jan 29 16:29:52 crc kubenswrapper[4714]: I0129 16:29:52.409571 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Jan 29 16:29:52 crc kubenswrapper[4714]: I0129 16:29:52.410445 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-2" podUID="869a701a-040e-44ea-98cc-53eb5f33c933" containerName="cinder-api-log" containerID="cri-o://fe14e97e807f7bfc78569e0b132f84ed0512e6e44920a45bdbbf02eab4f1fb0d" gracePeriod=30 Jan 29 16:29:52 crc kubenswrapper[4714]: I0129 16:29:52.410582 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-2" podUID="869a701a-040e-44ea-98cc-53eb5f33c933" containerName="cinder-api" containerID="cri-o://ee0d2c74772fa7181a1d5113e251172e052dd7202e52d6c038b2533ca027dac1" gracePeriod=30 Jan 29 16:29:52 crc kubenswrapper[4714]: I0129 16:29:52.418586 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Jan 29 16:29:52 crc kubenswrapper[4714]: I0129 16:29:52.418996 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-1" podUID="84becd41-a7aa-4b36-a6a7-8516c2e64909" containerName="cinder-api-log" containerID="cri-o://df6868e9d7ac6b0ebfef8c6a4c95fca607fa88217512148f0e9154c25064c11e" gracePeriod=30 Jan 29 16:29:52 crc kubenswrapper[4714]: I0129 16:29:52.419118 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-1" podUID="84becd41-a7aa-4b36-a6a7-8516c2e64909" containerName="cinder-api" containerID="cri-o://59d26536c5d18af3cf0b3379093a52042043c35adcb85fba86b4fe5fa5b56cb8" gracePeriod=30 Jan 29 16:29:52 crc kubenswrapper[4714]: I0129 16:29:52.425379 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="cinder-kuttl-tests/cinder-api-2" podUID="869a701a-040e-44ea-98cc-53eb5f33c933" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.100:8776/healthcheck\": EOF" Jan 29 16:29:52 crc kubenswrapper[4714]: I0129 16:29:52.430754 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="cinder-kuttl-tests/cinder-api-1" podUID="84becd41-a7aa-4b36-a6a7-8516c2e64909" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.99:8776/healthcheck\": EOF" Jan 29 16:29:53 crc kubenswrapper[4714]: I0129 16:29:53.039512 4714 generic.go:334] "Generic (PLEG): container finished" podID="84becd41-a7aa-4b36-a6a7-8516c2e64909" containerID="df6868e9d7ac6b0ebfef8c6a4c95fca607fa88217512148f0e9154c25064c11e" exitCode=143 Jan 29 16:29:53 crc kubenswrapper[4714]: I0129 16:29:53.039588 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"84becd41-a7aa-4b36-a6a7-8516c2e64909","Type":"ContainerDied","Data":"df6868e9d7ac6b0ebfef8c6a4c95fca607fa88217512148f0e9154c25064c11e"} Jan 29 16:29:53 crc kubenswrapper[4714]: I0129 16:29:53.043531 4714 generic.go:334] "Generic (PLEG): container finished" podID="869a701a-040e-44ea-98cc-53eb5f33c933" containerID="fe14e97e807f7bfc78569e0b132f84ed0512e6e44920a45bdbbf02eab4f1fb0d" exitCode=143 Jan 29 16:29:53 crc kubenswrapper[4714]: I0129 16:29:53.043570 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"869a701a-040e-44ea-98cc-53eb5f33c933","Type":"ContainerDied","Data":"fe14e97e807f7bfc78569e0b132f84ed0512e6e44920a45bdbbf02eab4f1fb0d"} Jan 29 16:29:56 crc kubenswrapper[4714]: I0129 16:29:56.856313 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/cinder-api-2" podUID="869a701a-040e-44ea-98cc-53eb5f33c933" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.100:8776/healthcheck\": read tcp 10.217.0.2:57768->10.217.0.100:8776: read: connection reset by peer" Jan 29 16:29:56 crc kubenswrapper[4714]: I0129 16:29:56.863983 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/cinder-api-1" podUID="84becd41-a7aa-4b36-a6a7-8516c2e64909" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.99:8776/healthcheck\": read tcp 10.217.0.2:57046->10.217.0.99:8776: read: connection reset by peer" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.080021 4714 generic.go:334] "Generic (PLEG): container finished" podID="869a701a-040e-44ea-98cc-53eb5f33c933" containerID="ee0d2c74772fa7181a1d5113e251172e052dd7202e52d6c038b2533ca027dac1" exitCode=0 Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.080203 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"869a701a-040e-44ea-98cc-53eb5f33c933","Type":"ContainerDied","Data":"ee0d2c74772fa7181a1d5113e251172e052dd7202e52d6c038b2533ca027dac1"} Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.082734 4714 generic.go:334] "Generic (PLEG): container finished" podID="84becd41-a7aa-4b36-a6a7-8516c2e64909" containerID="59d26536c5d18af3cf0b3379093a52042043c35adcb85fba86b4fe5fa5b56cb8" exitCode=0 Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.082764 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"84becd41-a7aa-4b36-a6a7-8516c2e64909","Type":"ContainerDied","Data":"59d26536c5d18af3cf0b3379093a52042043c35adcb85fba86b4fe5fa5b56cb8"} Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.275993 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.281644 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.398482 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84becd41-a7aa-4b36-a6a7-8516c2e64909-config-data\") pod \"84becd41-a7aa-4b36-a6a7-8516c2e64909\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.398536 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/869a701a-040e-44ea-98cc-53eb5f33c933-config-data\") pod \"869a701a-040e-44ea-98cc-53eb5f33c933\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.398567 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xk7jf\" (UniqueName: \"kubernetes.io/projected/84becd41-a7aa-4b36-a6a7-8516c2e64909-kube-api-access-xk7jf\") pod \"84becd41-a7aa-4b36-a6a7-8516c2e64909\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.398614 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/869a701a-040e-44ea-98cc-53eb5f33c933-etc-machine-id\") pod \"869a701a-040e-44ea-98cc-53eb5f33c933\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.398650 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zvnq\" (UniqueName: \"kubernetes.io/projected/869a701a-040e-44ea-98cc-53eb5f33c933-kube-api-access-4zvnq\") pod \"869a701a-040e-44ea-98cc-53eb5f33c933\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.398673 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84becd41-a7aa-4b36-a6a7-8516c2e64909-logs\") pod \"84becd41-a7aa-4b36-a6a7-8516c2e64909\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.398689 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/869a701a-040e-44ea-98cc-53eb5f33c933-logs\") pod \"869a701a-040e-44ea-98cc-53eb5f33c933\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.398748 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/869a701a-040e-44ea-98cc-53eb5f33c933-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "869a701a-040e-44ea-98cc-53eb5f33c933" (UID: "869a701a-040e-44ea-98cc-53eb5f33c933"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.398793 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84becd41-a7aa-4b36-a6a7-8516c2e64909-scripts\") pod \"84becd41-a7aa-4b36-a6a7-8516c2e64909\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.398870 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/869a701a-040e-44ea-98cc-53eb5f33c933-scripts\") pod \"869a701a-040e-44ea-98cc-53eb5f33c933\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.399303 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/84becd41-a7aa-4b36-a6a7-8516c2e64909-etc-machine-id\") pod \"84becd41-a7aa-4b36-a6a7-8516c2e64909\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.399336 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/869a701a-040e-44ea-98cc-53eb5f33c933-logs" (OuterVolumeSpecName: "logs") pod "869a701a-040e-44ea-98cc-53eb5f33c933" (UID: "869a701a-040e-44ea-98cc-53eb5f33c933"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.399346 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/869a701a-040e-44ea-98cc-53eb5f33c933-config-data-custom\") pod \"869a701a-040e-44ea-98cc-53eb5f33c933\" (UID: \"869a701a-040e-44ea-98cc-53eb5f33c933\") " Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.399332 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84becd41-a7aa-4b36-a6a7-8516c2e64909-logs" (OuterVolumeSpecName: "logs") pod "84becd41-a7aa-4b36-a6a7-8516c2e64909" (UID: "84becd41-a7aa-4b36-a6a7-8516c2e64909"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.399367 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/84becd41-a7aa-4b36-a6a7-8516c2e64909-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "84becd41-a7aa-4b36-a6a7-8516c2e64909" (UID: "84becd41-a7aa-4b36-a6a7-8516c2e64909"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.399390 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84becd41-a7aa-4b36-a6a7-8516c2e64909-config-data-custom\") pod \"84becd41-a7aa-4b36-a6a7-8516c2e64909\" (UID: \"84becd41-a7aa-4b36-a6a7-8516c2e64909\") " Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.399692 4714 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/84becd41-a7aa-4b36-a6a7-8516c2e64909-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.399707 4714 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/869a701a-040e-44ea-98cc-53eb5f33c933-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.399717 4714 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84becd41-a7aa-4b36-a6a7-8516c2e64909-logs\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.399727 4714 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/869a701a-040e-44ea-98cc-53eb5f33c933-logs\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.408300 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/869a701a-040e-44ea-98cc-53eb5f33c933-scripts" (OuterVolumeSpecName: "scripts") pod "869a701a-040e-44ea-98cc-53eb5f33c933" (UID: "869a701a-040e-44ea-98cc-53eb5f33c933"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.408374 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84becd41-a7aa-4b36-a6a7-8516c2e64909-scripts" (OuterVolumeSpecName: "scripts") pod "84becd41-a7aa-4b36-a6a7-8516c2e64909" (UID: "84becd41-a7aa-4b36-a6a7-8516c2e64909"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.409876 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/869a701a-040e-44ea-98cc-53eb5f33c933-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "869a701a-040e-44ea-98cc-53eb5f33c933" (UID: "869a701a-040e-44ea-98cc-53eb5f33c933"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.414239 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/869a701a-040e-44ea-98cc-53eb5f33c933-kube-api-access-4zvnq" (OuterVolumeSpecName: "kube-api-access-4zvnq") pod "869a701a-040e-44ea-98cc-53eb5f33c933" (UID: "869a701a-040e-44ea-98cc-53eb5f33c933"). InnerVolumeSpecName "kube-api-access-4zvnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.415496 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84becd41-a7aa-4b36-a6a7-8516c2e64909-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "84becd41-a7aa-4b36-a6a7-8516c2e64909" (UID: "84becd41-a7aa-4b36-a6a7-8516c2e64909"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.418032 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84becd41-a7aa-4b36-a6a7-8516c2e64909-kube-api-access-xk7jf" (OuterVolumeSpecName: "kube-api-access-xk7jf") pod "84becd41-a7aa-4b36-a6a7-8516c2e64909" (UID: "84becd41-a7aa-4b36-a6a7-8516c2e64909"). InnerVolumeSpecName "kube-api-access-xk7jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.443932 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/869a701a-040e-44ea-98cc-53eb5f33c933-config-data" (OuterVolumeSpecName: "config-data") pod "869a701a-040e-44ea-98cc-53eb5f33c933" (UID: "869a701a-040e-44ea-98cc-53eb5f33c933"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.463234 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84becd41-a7aa-4b36-a6a7-8516c2e64909-config-data" (OuterVolumeSpecName: "config-data") pod "84becd41-a7aa-4b36-a6a7-8516c2e64909" (UID: "84becd41-a7aa-4b36-a6a7-8516c2e64909"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.500925 4714 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/869a701a-040e-44ea-98cc-53eb5f33c933-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.501195 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/869a701a-040e-44ea-98cc-53eb5f33c933-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.501279 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84becd41-a7aa-4b36-a6a7-8516c2e64909-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.501359 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84becd41-a7aa-4b36-a6a7-8516c2e64909-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.501435 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/869a701a-040e-44ea-98cc-53eb5f33c933-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.501503 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xk7jf\" (UniqueName: \"kubernetes.io/projected/84becd41-a7aa-4b36-a6a7-8516c2e64909-kube-api-access-xk7jf\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.501582 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zvnq\" (UniqueName: \"kubernetes.io/projected/869a701a-040e-44ea-98cc-53eb5f33c933-kube-api-access-4zvnq\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.501653 4714 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84becd41-a7aa-4b36-a6a7-8516c2e64909-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.844172 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:29:57 crc kubenswrapper[4714]: I0129 16:29:57.844663 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.090768 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"869a701a-040e-44ea-98cc-53eb5f33c933","Type":"ContainerDied","Data":"2504822ce0a99b2d0cc7db311f85a30c6e6f161cbc200d841e855f4f3eb1ff5b"} Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.090814 4714 scope.go:117] "RemoveContainer" containerID="ee0d2c74772fa7181a1d5113e251172e052dd7202e52d6c038b2533ca027dac1" Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.090877 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-2" Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.094097 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"84becd41-a7aa-4b36-a6a7-8516c2e64909","Type":"ContainerDied","Data":"7f9cbec8d1d3b10670a11a02aa0992c33a6443ac29d99038a22ff077c3971712"} Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.094151 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-1" Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.109457 4714 scope.go:117] "RemoveContainer" containerID="fe14e97e807f7bfc78569e0b132f84ed0512e6e44920a45bdbbf02eab4f1fb0d" Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.133534 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.133901 4714 scope.go:117] "RemoveContainer" containerID="59d26536c5d18af3cf0b3379093a52042043c35adcb85fba86b4fe5fa5b56cb8" Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.149190 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.156621 4714 scope.go:117] "RemoveContainer" containerID="df6868e9d7ac6b0ebfef8c6a4c95fca607fa88217512148f0e9154c25064c11e" Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.156930 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.164593 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.191870 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84becd41-a7aa-4b36-a6a7-8516c2e64909" path="/var/lib/kubelet/pods/84becd41-a7aa-4b36-a6a7-8516c2e64909/volumes" Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.192613 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="869a701a-040e-44ea-98cc-53eb5f33c933" path="/var/lib/kubelet/pods/869a701a-040e-44ea-98cc-53eb5f33c933/volumes" Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.740091 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-9pvrg"] Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.752515 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-9pvrg"] Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.759682 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.759916 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="c0390b29-ac12-4c76-a954-8c7236d81661" containerName="cinder-scheduler" containerID="cri-o://8b1dd6490cae9482eb69b08b1702b33a571bfe8525bbdf572a2f8f5feb7f4e29" gracePeriod=30 Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.760031 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="c0390b29-ac12-4c76-a954-8c7236d81661" containerName="probe" containerID="cri-o://a3a659f2d53d2c5e83b9c4601491f3c732c7a7d0fe867de849d8aab7c8e4a69c" gracePeriod=30 Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.773197 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.773492 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="e01c83c7-65ba-4f1b-9d17-ba5a824216bb" containerName="cinder-backup" containerID="cri-o://f457144c29e9469b286400011d4bf2e3e4f7a3f73d20a9a7750c293dc9d6911b" gracePeriod=30 Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.773637 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="e01c83c7-65ba-4f1b-9d17-ba5a824216bb" containerName="probe" containerID="cri-o://e1bfab0c04a18d8e21e126e8983546eaa8cd1254da1312d48182dae10e5944bc" gracePeriod=30 Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.804013 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.848449 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.848716 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="c5a3f592-54c6-44f1-9f09-f366502287a6" containerName="cinder-api-log" containerID="cri-o://32d98c19e9cd977d60b8a6256b941e6d21b55abe95ffb10609d846a79c267c86" gracePeriod=30 Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.849149 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="c5a3f592-54c6-44f1-9f09-f366502287a6" containerName="cinder-api" containerID="cri-o://922b00d2d16cf09fabe14b15b7a7648c66244d5a615d55c77b2cc333c2095cf3" gracePeriod=30 Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.855243 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder9de3-account-delete-6gpgf"] Jan 29 16:29:58 crc kubenswrapper[4714]: E0129 16:29:58.855868 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="869a701a-040e-44ea-98cc-53eb5f33c933" containerName="cinder-api" Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.855888 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="869a701a-040e-44ea-98cc-53eb5f33c933" containerName="cinder-api" Jan 29 16:29:58 crc kubenswrapper[4714]: E0129 16:29:58.855909 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84becd41-a7aa-4b36-a6a7-8516c2e64909" containerName="cinder-api" Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.855917 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="84becd41-a7aa-4b36-a6a7-8516c2e64909" containerName="cinder-api" Jan 29 16:29:58 crc kubenswrapper[4714]: E0129 16:29:58.855931 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84becd41-a7aa-4b36-a6a7-8516c2e64909" containerName="cinder-api-log" Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.856015 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="84becd41-a7aa-4b36-a6a7-8516c2e64909" containerName="cinder-api-log" Jan 29 16:29:58 crc kubenswrapper[4714]: E0129 16:29:58.856032 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="869a701a-040e-44ea-98cc-53eb5f33c933" containerName="cinder-api-log" Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.856039 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="869a701a-040e-44ea-98cc-53eb5f33c933" containerName="cinder-api-log" Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.856199 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="84becd41-a7aa-4b36-a6a7-8516c2e64909" containerName="cinder-api-log" Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.856221 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="84becd41-a7aa-4b36-a6a7-8516c2e64909" containerName="cinder-api" Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.856232 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="869a701a-040e-44ea-98cc-53eb5f33c933" containerName="cinder-api" Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.856245 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="869a701a-040e-44ea-98cc-53eb5f33c933" containerName="cinder-api-log" Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.856769 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder9de3-account-delete-6gpgf" Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.871119 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder9de3-account-delete-6gpgf"] Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.927130 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d989387-6924-48df-a1d4-1c63911dd476-operator-scripts\") pod \"cinder9de3-account-delete-6gpgf\" (UID: \"8d989387-6924-48df-a1d4-1c63911dd476\") " pod="cinder-kuttl-tests/cinder9de3-account-delete-6gpgf" Jan 29 16:29:58 crc kubenswrapper[4714]: I0129 16:29:58.927181 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftsmf\" (UniqueName: \"kubernetes.io/projected/8d989387-6924-48df-a1d4-1c63911dd476-kube-api-access-ftsmf\") pod \"cinder9de3-account-delete-6gpgf\" (UID: \"8d989387-6924-48df-a1d4-1c63911dd476\") " pod="cinder-kuttl-tests/cinder9de3-account-delete-6gpgf" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.029031 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d989387-6924-48df-a1d4-1c63911dd476-operator-scripts\") pod \"cinder9de3-account-delete-6gpgf\" (UID: \"8d989387-6924-48df-a1d4-1c63911dd476\") " pod="cinder-kuttl-tests/cinder9de3-account-delete-6gpgf" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.029095 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftsmf\" (UniqueName: \"kubernetes.io/projected/8d989387-6924-48df-a1d4-1c63911dd476-kube-api-access-ftsmf\") pod \"cinder9de3-account-delete-6gpgf\" (UID: \"8d989387-6924-48df-a1d4-1c63911dd476\") " pod="cinder-kuttl-tests/cinder9de3-account-delete-6gpgf" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.029890 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d989387-6924-48df-a1d4-1c63911dd476-operator-scripts\") pod \"cinder9de3-account-delete-6gpgf\" (UID: \"8d989387-6924-48df-a1d4-1c63911dd476\") " pod="cinder-kuttl-tests/cinder9de3-account-delete-6gpgf" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.055686 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftsmf\" (UniqueName: \"kubernetes.io/projected/8d989387-6924-48df-a1d4-1c63911dd476-kube-api-access-ftsmf\") pod \"cinder9de3-account-delete-6gpgf\" (UID: \"8d989387-6924-48df-a1d4-1c63911dd476\") " pod="cinder-kuttl-tests/cinder9de3-account-delete-6gpgf" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.102183 4714 generic.go:334] "Generic (PLEG): container finished" podID="c5a3f592-54c6-44f1-9f09-f366502287a6" containerID="32d98c19e9cd977d60b8a6256b941e6d21b55abe95ffb10609d846a79c267c86" exitCode=143 Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.102231 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"c5a3f592-54c6-44f1-9f09-f366502287a6","Type":"ContainerDied","Data":"32d98c19e9cd977d60b8a6256b941e6d21b55abe95ffb10609d846a79c267c86"} Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.104291 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"2e154d80-4b79-4f74-809e-c1c274ed4063","Type":"ContainerDied","Data":"aa9a0a1ce04cf5e776943ebe4b4ffc887cc7f0d7111f44e2e55f143d9edbcb9b"} Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.104322 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa9a0a1ce04cf5e776943ebe4b4ffc887cc7f0d7111f44e2e55f143d9edbcb9b" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.106772 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.177766 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder9de3-account-delete-6gpgf" Jan 29 16:29:59 crc kubenswrapper[4714]: E0129 16:29:59.188328 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231240 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-var-locks-cinder\") pod \"2e154d80-4b79-4f74-809e-c1c274ed4063\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231303 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-etc-machine-id\") pod \"2e154d80-4b79-4f74-809e-c1c274ed4063\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231347 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e154d80-4b79-4f74-809e-c1c274ed4063-config-data\") pod \"2e154d80-4b79-4f74-809e-c1c274ed4063\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231376 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-dev\") pod \"2e154d80-4b79-4f74-809e-c1c274ed4063\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231389 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "2e154d80-4b79-4f74-809e-c1c274ed4063" (UID: "2e154d80-4b79-4f74-809e-c1c274ed4063"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231420 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-etc-nvme\") pod \"2e154d80-4b79-4f74-809e-c1c274ed4063\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231444 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2e154d80-4b79-4f74-809e-c1c274ed4063" (UID: "2e154d80-4b79-4f74-809e-c1c274ed4063"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231463 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-sys\") pod \"2e154d80-4b79-4f74-809e-c1c274ed4063\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231465 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-dev" (OuterVolumeSpecName: "dev") pod "2e154d80-4b79-4f74-809e-c1c274ed4063" (UID: "2e154d80-4b79-4f74-809e-c1c274ed4063"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231486 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-sys" (OuterVolumeSpecName: "sys") pod "2e154d80-4b79-4f74-809e-c1c274ed4063" (UID: "2e154d80-4b79-4f74-809e-c1c274ed4063"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231488 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "2e154d80-4b79-4f74-809e-c1c274ed4063" (UID: "2e154d80-4b79-4f74-809e-c1c274ed4063"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231501 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-lib-modules\") pod \"2e154d80-4b79-4f74-809e-c1c274ed4063\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231564 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "2e154d80-4b79-4f74-809e-c1c274ed4063" (UID: "2e154d80-4b79-4f74-809e-c1c274ed4063"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231593 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-var-locks-brick\") pod \"2e154d80-4b79-4f74-809e-c1c274ed4063\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231617 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e154d80-4b79-4f74-809e-c1c274ed4063-scripts\") pod \"2e154d80-4b79-4f74-809e-c1c274ed4063\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231620 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "2e154d80-4b79-4f74-809e-c1c274ed4063" (UID: "2e154d80-4b79-4f74-809e-c1c274ed4063"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231661 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbcbm\" (UniqueName: \"kubernetes.io/projected/2e154d80-4b79-4f74-809e-c1c274ed4063-kube-api-access-dbcbm\") pod \"2e154d80-4b79-4f74-809e-c1c274ed4063\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231709 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-var-lib-cinder\") pod \"2e154d80-4b79-4f74-809e-c1c274ed4063\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231730 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e154d80-4b79-4f74-809e-c1c274ed4063-config-data-custom\") pod \"2e154d80-4b79-4f74-809e-c1c274ed4063\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231794 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-etc-iscsi\") pod \"2e154d80-4b79-4f74-809e-c1c274ed4063\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231816 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-run\") pod \"2e154d80-4b79-4f74-809e-c1c274ed4063\" (UID: \"2e154d80-4b79-4f74-809e-c1c274ed4063\") " Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.231833 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "2e154d80-4b79-4f74-809e-c1c274ed4063" (UID: "2e154d80-4b79-4f74-809e-c1c274ed4063"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.232241 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "2e154d80-4b79-4f74-809e-c1c274ed4063" (UID: "2e154d80-4b79-4f74-809e-c1c274ed4063"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.232272 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-run" (OuterVolumeSpecName: "run") pod "2e154d80-4b79-4f74-809e-c1c274ed4063" (UID: "2e154d80-4b79-4f74-809e-c1c274ed4063"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.232395 4714 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.232410 4714 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-etc-iscsi\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.232420 4714 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-run\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.232432 4714 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.232446 4714 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.232458 4714 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-dev\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.232469 4714 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-etc-nvme\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.232480 4714 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-sys\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.232490 4714 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-lib-modules\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.232500 4714 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2e154d80-4b79-4f74-809e-c1c274ed4063-var-locks-brick\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.238431 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e154d80-4b79-4f74-809e-c1c274ed4063-scripts" (OuterVolumeSpecName: "scripts") pod "2e154d80-4b79-4f74-809e-c1c274ed4063" (UID: "2e154d80-4b79-4f74-809e-c1c274ed4063"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.239243 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e154d80-4b79-4f74-809e-c1c274ed4063-kube-api-access-dbcbm" (OuterVolumeSpecName: "kube-api-access-dbcbm") pod "2e154d80-4b79-4f74-809e-c1c274ed4063" (UID: "2e154d80-4b79-4f74-809e-c1c274ed4063"). InnerVolumeSpecName "kube-api-access-dbcbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.240621 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e154d80-4b79-4f74-809e-c1c274ed4063-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2e154d80-4b79-4f74-809e-c1c274ed4063" (UID: "2e154d80-4b79-4f74-809e-c1c274ed4063"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.309496 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e154d80-4b79-4f74-809e-c1c274ed4063-config-data" (OuterVolumeSpecName: "config-data") pod "2e154d80-4b79-4f74-809e-c1c274ed4063" (UID: "2e154d80-4b79-4f74-809e-c1c274ed4063"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.336574 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e154d80-4b79-4f74-809e-c1c274ed4063-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.336600 4714 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e154d80-4b79-4f74-809e-c1c274ed4063-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.336612 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbcbm\" (UniqueName: \"kubernetes.io/projected/2e154d80-4b79-4f74-809e-c1c274ed4063-kube-api-access-dbcbm\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.336653 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e154d80-4b79-4f74-809e-c1c274ed4063-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 29 16:29:59 crc kubenswrapper[4714]: I0129 16:29:59.592548 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder9de3-account-delete-6gpgf"] Jan 29 16:29:59 crc kubenswrapper[4714]: W0129 16:29:59.596698 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d989387_6924_48df_a1d4_1c63911dd476.slice/crio-16c8423dc7a88554a04643282ed7d3e92befa98e3c24c9a8e1f726e72cddc134 WatchSource:0}: Error finding container 16c8423dc7a88554a04643282ed7d3e92befa98e3c24c9a8e1f726e72cddc134: Status 404 returned error can't find the container with id 16c8423dc7a88554a04643282ed7d3e92befa98e3c24c9a8e1f726e72cddc134 Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.114313 4714 generic.go:334] "Generic (PLEG): container finished" podID="c0390b29-ac12-4c76-a954-8c7236d81661" containerID="a3a659f2d53d2c5e83b9c4601491f3c732c7a7d0fe867de849d8aab7c8e4a69c" exitCode=0 Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.114379 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"c0390b29-ac12-4c76-a954-8c7236d81661","Type":"ContainerDied","Data":"a3a659f2d53d2c5e83b9c4601491f3c732c7a7d0fe867de849d8aab7c8e4a69c"} Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.116376 4714 generic.go:334] "Generic (PLEG): container finished" podID="e01c83c7-65ba-4f1b-9d17-ba5a824216bb" containerID="e1bfab0c04a18d8e21e126e8983546eaa8cd1254da1312d48182dae10e5944bc" exitCode=0 Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.116425 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"e01c83c7-65ba-4f1b-9d17-ba5a824216bb","Type":"ContainerDied","Data":"e1bfab0c04a18d8e21e126e8983546eaa8cd1254da1312d48182dae10e5944bc"} Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.117968 4714 generic.go:334] "Generic (PLEG): container finished" podID="8d989387-6924-48df-a1d4-1c63911dd476" containerID="b17d360d1529ce324f10ea0628f0cee292edf28a50e5d75cc7f2606e49a8da6e" exitCode=0 Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.118046 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.118140 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder9de3-account-delete-6gpgf" event={"ID":"8d989387-6924-48df-a1d4-1c63911dd476","Type":"ContainerDied","Data":"b17d360d1529ce324f10ea0628f0cee292edf28a50e5d75cc7f2606e49a8da6e"} Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.118184 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder9de3-account-delete-6gpgf" event={"ID":"8d989387-6924-48df-a1d4-1c63911dd476","Type":"ContainerStarted","Data":"16c8423dc7a88554a04643282ed7d3e92befa98e3c24c9a8e1f726e72cddc134"} Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.152362 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29495070-b47jt"] Jan 29 16:30:00 crc kubenswrapper[4714]: E0129 16:30:00.152666 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="probe" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.152680 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="probe" Jan 29 16:30:00 crc kubenswrapper[4714]: E0129 16:30:00.152688 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="cinder-volume" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.152694 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="cinder-volume" Jan 29 16:30:00 crc kubenswrapper[4714]: E0129 16:30:00.152703 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="cinder-volume" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.152708 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="cinder-volume" Jan 29 16:30:00 crc kubenswrapper[4714]: E0129 16:30:00.152730 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="cinder-volume" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.152738 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="cinder-volume" Jan 29 16:30:00 crc kubenswrapper[4714]: E0129 16:30:00.152748 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="probe" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.152754 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="probe" Jan 29 16:30:00 crc kubenswrapper[4714]: E0129 16:30:00.152763 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="cinder-volume" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.152768 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="cinder-volume" Jan 29 16:30:00 crc kubenswrapper[4714]: E0129 16:30:00.152779 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="probe" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.152785 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="probe" Jan 29 16:30:00 crc kubenswrapper[4714]: E0129 16:30:00.152796 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="probe" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.152802 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="probe" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.152927 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="probe" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.152951 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="probe" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.152962 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="cinder-volume" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.152969 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="cinder-volume" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.152977 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="probe" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.152986 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="cinder-volume" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.153423 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495070-b47jt" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.160330 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.160353 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.167145 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29495070-b47jt"] Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.196771 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f09789c2-52ed-4321-95f8-02c3b3f271e3" path="/var/lib/kubelet/pods/f09789c2-52ed-4321-95f8-02c3b3f271e3/volumes" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.201321 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.205543 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.253854 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76e30b28-5660-4c1a-a31b-626cc3bb6c38-config-volume\") pod \"collect-profiles-29495070-b47jt\" (UID: \"76e30b28-5660-4c1a-a31b-626cc3bb6c38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495070-b47jt" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.253909 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76e30b28-5660-4c1a-a31b-626cc3bb6c38-secret-volume\") pod \"collect-profiles-29495070-b47jt\" (UID: \"76e30b28-5660-4c1a-a31b-626cc3bb6c38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495070-b47jt" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.253979 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqqvq\" (UniqueName: \"kubernetes.io/projected/76e30b28-5660-4c1a-a31b-626cc3bb6c38-kube-api-access-nqqvq\") pod \"collect-profiles-29495070-b47jt\" (UID: \"76e30b28-5660-4c1a-a31b-626cc3bb6c38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495070-b47jt" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.355029 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76e30b28-5660-4c1a-a31b-626cc3bb6c38-config-volume\") pod \"collect-profiles-29495070-b47jt\" (UID: \"76e30b28-5660-4c1a-a31b-626cc3bb6c38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495070-b47jt" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.355313 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76e30b28-5660-4c1a-a31b-626cc3bb6c38-secret-volume\") pod \"collect-profiles-29495070-b47jt\" (UID: \"76e30b28-5660-4c1a-a31b-626cc3bb6c38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495070-b47jt" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.355346 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqqvq\" (UniqueName: \"kubernetes.io/projected/76e30b28-5660-4c1a-a31b-626cc3bb6c38-kube-api-access-nqqvq\") pod \"collect-profiles-29495070-b47jt\" (UID: \"76e30b28-5660-4c1a-a31b-626cc3bb6c38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495070-b47jt" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.356437 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76e30b28-5660-4c1a-a31b-626cc3bb6c38-config-volume\") pod \"collect-profiles-29495070-b47jt\" (UID: \"76e30b28-5660-4c1a-a31b-626cc3bb6c38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495070-b47jt" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.361973 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76e30b28-5660-4c1a-a31b-626cc3bb6c38-secret-volume\") pod \"collect-profiles-29495070-b47jt\" (UID: \"76e30b28-5660-4c1a-a31b-626cc3bb6c38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495070-b47jt" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.373288 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqqvq\" (UniqueName: \"kubernetes.io/projected/76e30b28-5660-4c1a-a31b-626cc3bb6c38-kube-api-access-nqqvq\") pod \"collect-profiles-29495070-b47jt\" (UID: \"76e30b28-5660-4c1a-a31b-626cc3bb6c38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495070-b47jt" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.505784 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495070-b47jt" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.629230 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.725340 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29495070-b47jt"] Jan 29 16:30:00 crc kubenswrapper[4714]: W0129 16:30:00.728076 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76e30b28_5660_4c1a_a31b_626cc3bb6c38.slice/crio-fcb9acc88fce040f1e751e565b0e7e67527c868c890f2f7bba33c01e544174b1 WatchSource:0}: Error finding container fcb9acc88fce040f1e751e565b0e7e67527c868c890f2f7bba33c01e544174b1: Status 404 returned error can't find the container with id fcb9acc88fce040f1e751e565b0e7e67527c868c890f2f7bba33c01e544174b1 Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761259 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-var-lib-cinder\") pod \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761325 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "e01c83c7-65ba-4f1b-9d17-ba5a824216bb" (UID: "e01c83c7-65ba-4f1b-9d17-ba5a824216bb"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761338 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-lib-modules\") pod \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761365 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "e01c83c7-65ba-4f1b-9d17-ba5a824216bb" (UID: "e01c83c7-65ba-4f1b-9d17-ba5a824216bb"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761388 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-var-locks-brick\") pod \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761432 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-scripts\") pod \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761464 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "e01c83c7-65ba-4f1b-9d17-ba5a824216bb" (UID: "e01c83c7-65ba-4f1b-9d17-ba5a824216bb"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761471 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5dd5\" (UniqueName: \"kubernetes.io/projected/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-kube-api-access-s5dd5\") pod \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761500 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-config-data\") pod \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761525 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-config-data-custom\") pod \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761557 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-etc-nvme\") pod \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761577 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-etc-iscsi\") pod \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761670 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-run\") pod \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761691 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-var-locks-cinder\") pod \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761661 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "e01c83c7-65ba-4f1b-9d17-ba5a824216bb" (UID: "e01c83c7-65ba-4f1b-9d17-ba5a824216bb"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761695 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "e01c83c7-65ba-4f1b-9d17-ba5a824216bb" (UID: "e01c83c7-65ba-4f1b-9d17-ba5a824216bb"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761719 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-run" (OuterVolumeSpecName: "run") pod "e01c83c7-65ba-4f1b-9d17-ba5a824216bb" (UID: "e01c83c7-65ba-4f1b-9d17-ba5a824216bb"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761780 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-dev\") pod \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761812 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "e01c83c7-65ba-4f1b-9d17-ba5a824216bb" (UID: "e01c83c7-65ba-4f1b-9d17-ba5a824216bb"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761834 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-sys\") pod \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761849 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-dev" (OuterVolumeSpecName: "dev") pod "e01c83c7-65ba-4f1b-9d17-ba5a824216bb" (UID: "e01c83c7-65ba-4f1b-9d17-ba5a824216bb"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761906 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-etc-machine-id\") pod \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\" (UID: \"e01c83c7-65ba-4f1b-9d17-ba5a824216bb\") " Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.761957 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-sys" (OuterVolumeSpecName: "sys") pod "e01c83c7-65ba-4f1b-9d17-ba5a824216bb" (UID: "e01c83c7-65ba-4f1b-9d17-ba5a824216bb"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.762037 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e01c83c7-65ba-4f1b-9d17-ba5a824216bb" (UID: "e01c83c7-65ba-4f1b-9d17-ba5a824216bb"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.762302 4714 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-etc-nvme\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.762326 4714 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-etc-iscsi\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.762340 4714 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.762354 4714 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-run\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.762364 4714 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-dev\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.762375 4714 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-sys\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.762387 4714 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.762399 4714 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.762411 4714 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-lib-modules\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.762421 4714 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-var-locks-brick\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.764622 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-kube-api-access-s5dd5" (OuterVolumeSpecName: "kube-api-access-s5dd5") pod "e01c83c7-65ba-4f1b-9d17-ba5a824216bb" (UID: "e01c83c7-65ba-4f1b-9d17-ba5a824216bb"). InnerVolumeSpecName "kube-api-access-s5dd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.765130 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-scripts" (OuterVolumeSpecName: "scripts") pod "e01c83c7-65ba-4f1b-9d17-ba5a824216bb" (UID: "e01c83c7-65ba-4f1b-9d17-ba5a824216bb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.765326 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e01c83c7-65ba-4f1b-9d17-ba5a824216bb" (UID: "e01c83c7-65ba-4f1b-9d17-ba5a824216bb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.837392 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-config-data" (OuterVolumeSpecName: "config-data") pod "e01c83c7-65ba-4f1b-9d17-ba5a824216bb" (UID: "e01c83c7-65ba-4f1b-9d17-ba5a824216bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.864485 4714 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.864522 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5dd5\" (UniqueName: \"kubernetes.io/projected/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-kube-api-access-s5dd5\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.864535 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:00 crc kubenswrapper[4714]: I0129 16:30:00.864544 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e01c83c7-65ba-4f1b-9d17-ba5a824216bb-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.125380 4714 generic.go:334] "Generic (PLEG): container finished" podID="76e30b28-5660-4c1a-a31b-626cc3bb6c38" containerID="877ac8183fb77e185f572df964edcba4a95cca50385422071d4d1af26ee35620" exitCode=0 Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.125454 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29495070-b47jt" event={"ID":"76e30b28-5660-4c1a-a31b-626cc3bb6c38","Type":"ContainerDied","Data":"877ac8183fb77e185f572df964edcba4a95cca50385422071d4d1af26ee35620"} Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.125720 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29495070-b47jt" event={"ID":"76e30b28-5660-4c1a-a31b-626cc3bb6c38","Type":"ContainerStarted","Data":"fcb9acc88fce040f1e751e565b0e7e67527c868c890f2f7bba33c01e544174b1"} Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.132137 4714 generic.go:334] "Generic (PLEG): container finished" podID="e01c83c7-65ba-4f1b-9d17-ba5a824216bb" containerID="f457144c29e9469b286400011d4bf2e3e4f7a3f73d20a9a7750c293dc9d6911b" exitCode=0 Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.132460 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"e01c83c7-65ba-4f1b-9d17-ba5a824216bb","Type":"ContainerDied","Data":"f457144c29e9469b286400011d4bf2e3e4f7a3f73d20a9a7750c293dc9d6911b"} Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.132551 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"e01c83c7-65ba-4f1b-9d17-ba5a824216bb","Type":"ContainerDied","Data":"198a3df4efafc7e3421f9f093f38267c1562e37ed9290df232bcf8b82972d9a2"} Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.132578 4714 scope.go:117] "RemoveContainer" containerID="e1bfab0c04a18d8e21e126e8983546eaa8cd1254da1312d48182dae10e5944bc" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.132478 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.166447 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.169302 4714 scope.go:117] "RemoveContainer" containerID="f457144c29e9469b286400011d4bf2e3e4f7a3f73d20a9a7750c293dc9d6911b" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.171191 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.187494 4714 scope.go:117] "RemoveContainer" containerID="e1bfab0c04a18d8e21e126e8983546eaa8cd1254da1312d48182dae10e5944bc" Jan 29 16:30:01 crc kubenswrapper[4714]: E0129 16:30:01.187900 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1bfab0c04a18d8e21e126e8983546eaa8cd1254da1312d48182dae10e5944bc\": container with ID starting with e1bfab0c04a18d8e21e126e8983546eaa8cd1254da1312d48182dae10e5944bc not found: ID does not exist" containerID="e1bfab0c04a18d8e21e126e8983546eaa8cd1254da1312d48182dae10e5944bc" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.187942 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1bfab0c04a18d8e21e126e8983546eaa8cd1254da1312d48182dae10e5944bc"} err="failed to get container status \"e1bfab0c04a18d8e21e126e8983546eaa8cd1254da1312d48182dae10e5944bc\": rpc error: code = NotFound desc = could not find container \"e1bfab0c04a18d8e21e126e8983546eaa8cd1254da1312d48182dae10e5944bc\": container with ID starting with e1bfab0c04a18d8e21e126e8983546eaa8cd1254da1312d48182dae10e5944bc not found: ID does not exist" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.187965 4714 scope.go:117] "RemoveContainer" containerID="f457144c29e9469b286400011d4bf2e3e4f7a3f73d20a9a7750c293dc9d6911b" Jan 29 16:30:01 crc kubenswrapper[4714]: E0129 16:30:01.188294 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f457144c29e9469b286400011d4bf2e3e4f7a3f73d20a9a7750c293dc9d6911b\": container with ID starting with f457144c29e9469b286400011d4bf2e3e4f7a3f73d20a9a7750c293dc9d6911b not found: ID does not exist" containerID="f457144c29e9469b286400011d4bf2e3e4f7a3f73d20a9a7750c293dc9d6911b" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.188324 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f457144c29e9469b286400011d4bf2e3e4f7a3f73d20a9a7750c293dc9d6911b"} err="failed to get container status \"f457144c29e9469b286400011d4bf2e3e4f7a3f73d20a9a7750c293dc9d6911b\": rpc error: code = NotFound desc = could not find container \"f457144c29e9469b286400011d4bf2e3e4f7a3f73d20a9a7750c293dc9d6911b\": container with ID starting with f457144c29e9469b286400011d4bf2e3e4f7a3f73d20a9a7750c293dc9d6911b not found: ID does not exist" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.417523 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder9de3-account-delete-6gpgf" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.474098 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftsmf\" (UniqueName: \"kubernetes.io/projected/8d989387-6924-48df-a1d4-1c63911dd476-kube-api-access-ftsmf\") pod \"8d989387-6924-48df-a1d4-1c63911dd476\" (UID: \"8d989387-6924-48df-a1d4-1c63911dd476\") " Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.474189 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d989387-6924-48df-a1d4-1c63911dd476-operator-scripts\") pod \"8d989387-6924-48df-a1d4-1c63911dd476\" (UID: \"8d989387-6924-48df-a1d4-1c63911dd476\") " Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.474869 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d989387-6924-48df-a1d4-1c63911dd476-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8d989387-6924-48df-a1d4-1c63911dd476" (UID: "8d989387-6924-48df-a1d4-1c63911dd476"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.496065 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d989387-6924-48df-a1d4-1c63911dd476-kube-api-access-ftsmf" (OuterVolumeSpecName: "kube-api-access-ftsmf") pod "8d989387-6924-48df-a1d4-1c63911dd476" (UID: "8d989387-6924-48df-a1d4-1c63911dd476"). InnerVolumeSpecName "kube-api-access-ftsmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.576073 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftsmf\" (UniqueName: \"kubernetes.io/projected/8d989387-6924-48df-a1d4-1c63911dd476-kube-api-access-ftsmf\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.576110 4714 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d989387-6924-48df-a1d4-1c63911dd476-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.651487 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.777847 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rxqx\" (UniqueName: \"kubernetes.io/projected/c0390b29-ac12-4c76-a954-8c7236d81661-kube-api-access-6rxqx\") pod \"c0390b29-ac12-4c76-a954-8c7236d81661\" (UID: \"c0390b29-ac12-4c76-a954-8c7236d81661\") " Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.777900 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0390b29-ac12-4c76-a954-8c7236d81661-config-data\") pod \"c0390b29-ac12-4c76-a954-8c7236d81661\" (UID: \"c0390b29-ac12-4c76-a954-8c7236d81661\") " Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.778026 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0390b29-ac12-4c76-a954-8c7236d81661-config-data-custom\") pod \"c0390b29-ac12-4c76-a954-8c7236d81661\" (UID: \"c0390b29-ac12-4c76-a954-8c7236d81661\") " Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.778081 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0390b29-ac12-4c76-a954-8c7236d81661-scripts\") pod \"c0390b29-ac12-4c76-a954-8c7236d81661\" (UID: \"c0390b29-ac12-4c76-a954-8c7236d81661\") " Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.778114 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c0390b29-ac12-4c76-a954-8c7236d81661-etc-machine-id\") pod \"c0390b29-ac12-4c76-a954-8c7236d81661\" (UID: \"c0390b29-ac12-4c76-a954-8c7236d81661\") " Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.778351 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c0390b29-ac12-4c76-a954-8c7236d81661-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c0390b29-ac12-4c76-a954-8c7236d81661" (UID: "c0390b29-ac12-4c76-a954-8c7236d81661"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.780828 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0390b29-ac12-4c76-a954-8c7236d81661-kube-api-access-6rxqx" (OuterVolumeSpecName: "kube-api-access-6rxqx") pod "c0390b29-ac12-4c76-a954-8c7236d81661" (UID: "c0390b29-ac12-4c76-a954-8c7236d81661"). InnerVolumeSpecName "kube-api-access-6rxqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.781154 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0390b29-ac12-4c76-a954-8c7236d81661-scripts" (OuterVolumeSpecName: "scripts") pod "c0390b29-ac12-4c76-a954-8c7236d81661" (UID: "c0390b29-ac12-4c76-a954-8c7236d81661"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.782163 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0390b29-ac12-4c76-a954-8c7236d81661-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c0390b29-ac12-4c76-a954-8c7236d81661" (UID: "c0390b29-ac12-4c76-a954-8c7236d81661"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.849063 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0390b29-ac12-4c76-a954-8c7236d81661-config-data" (OuterVolumeSpecName: "config-data") pod "c0390b29-ac12-4c76-a954-8c7236d81661" (UID: "c0390b29-ac12-4c76-a954-8c7236d81661"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.879140 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0390b29-ac12-4c76-a954-8c7236d81661-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.879194 4714 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0390b29-ac12-4c76-a954-8c7236d81661-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.879208 4714 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c0390b29-ac12-4c76-a954-8c7236d81661-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.879221 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rxqx\" (UniqueName: \"kubernetes.io/projected/c0390b29-ac12-4c76-a954-8c7236d81661-kube-api-access-6rxqx\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:01 crc kubenswrapper[4714]: I0129 16:30:01.879235 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0390b29-ac12-4c76-a954-8c7236d81661-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.022184 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/cinder-api-0" podUID="c5a3f592-54c6-44f1-9f09-f366502287a6" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.98:8776/healthcheck\": read tcp 10.217.0.2:33032->10.217.0.98:8776: read: connection reset by peer" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.149579 4714 generic.go:334] "Generic (PLEG): container finished" podID="c5a3f592-54c6-44f1-9f09-f366502287a6" containerID="922b00d2d16cf09fabe14b15b7a7648c66244d5a615d55c77b2cc333c2095cf3" exitCode=0 Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.149684 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"c5a3f592-54c6-44f1-9f09-f366502287a6","Type":"ContainerDied","Data":"922b00d2d16cf09fabe14b15b7a7648c66244d5a615d55c77b2cc333c2095cf3"} Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.152333 4714 generic.go:334] "Generic (PLEG): container finished" podID="c0390b29-ac12-4c76-a954-8c7236d81661" containerID="8b1dd6490cae9482eb69b08b1702b33a571bfe8525bbdf572a2f8f5feb7f4e29" exitCode=0 Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.152398 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"c0390b29-ac12-4c76-a954-8c7236d81661","Type":"ContainerDied","Data":"8b1dd6490cae9482eb69b08b1702b33a571bfe8525bbdf572a2f8f5feb7f4e29"} Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.152397 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.152426 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"c0390b29-ac12-4c76-a954-8c7236d81661","Type":"ContainerDied","Data":"a7e3257c239df083499879bab78d2e91169e0bd8d92a8c5fac288924f3619908"} Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.152450 4714 scope.go:117] "RemoveContainer" containerID="a3a659f2d53d2c5e83b9c4601491f3c732c7a7d0fe867de849d8aab7c8e4a69c" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.156250 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder9de3-account-delete-6gpgf" event={"ID":"8d989387-6924-48df-a1d4-1c63911dd476","Type":"ContainerDied","Data":"16c8423dc7a88554a04643282ed7d3e92befa98e3c24c9a8e1f726e72cddc134"} Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.156286 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16c8423dc7a88554a04643282ed7d3e92befa98e3c24c9a8e1f726e72cddc134" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.156392 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder9de3-account-delete-6gpgf" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.210703 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" path="/var/lib/kubelet/pods/2e154d80-4b79-4f74-809e-c1c274ed4063/volumes" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.211473 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e01c83c7-65ba-4f1b-9d17-ba5a824216bb" path="/var/lib/kubelet/pods/e01c83c7-65ba-4f1b-9d17-ba5a824216bb/volumes" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.212057 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.212088 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.214920 4714 scope.go:117] "RemoveContainer" containerID="8b1dd6490cae9482eb69b08b1702b33a571bfe8525bbdf572a2f8f5feb7f4e29" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.236159 4714 scope.go:117] "RemoveContainer" containerID="a3a659f2d53d2c5e83b9c4601491f3c732c7a7d0fe867de849d8aab7c8e4a69c" Jan 29 16:30:02 crc kubenswrapper[4714]: E0129 16:30:02.238330 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3a659f2d53d2c5e83b9c4601491f3c732c7a7d0fe867de849d8aab7c8e4a69c\": container with ID starting with a3a659f2d53d2c5e83b9c4601491f3c732c7a7d0fe867de849d8aab7c8e4a69c not found: ID does not exist" containerID="a3a659f2d53d2c5e83b9c4601491f3c732c7a7d0fe867de849d8aab7c8e4a69c" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.238358 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3a659f2d53d2c5e83b9c4601491f3c732c7a7d0fe867de849d8aab7c8e4a69c"} err="failed to get container status \"a3a659f2d53d2c5e83b9c4601491f3c732c7a7d0fe867de849d8aab7c8e4a69c\": rpc error: code = NotFound desc = could not find container \"a3a659f2d53d2c5e83b9c4601491f3c732c7a7d0fe867de849d8aab7c8e4a69c\": container with ID starting with a3a659f2d53d2c5e83b9c4601491f3c732c7a7d0fe867de849d8aab7c8e4a69c not found: ID does not exist" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.238379 4714 scope.go:117] "RemoveContainer" containerID="8b1dd6490cae9482eb69b08b1702b33a571bfe8525bbdf572a2f8f5feb7f4e29" Jan 29 16:30:02 crc kubenswrapper[4714]: E0129 16:30:02.249238 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b1dd6490cae9482eb69b08b1702b33a571bfe8525bbdf572a2f8f5feb7f4e29\": container with ID starting with 8b1dd6490cae9482eb69b08b1702b33a571bfe8525bbdf572a2f8f5feb7f4e29 not found: ID does not exist" containerID="8b1dd6490cae9482eb69b08b1702b33a571bfe8525bbdf572a2f8f5feb7f4e29" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.249274 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b1dd6490cae9482eb69b08b1702b33a571bfe8525bbdf572a2f8f5feb7f4e29"} err="failed to get container status \"8b1dd6490cae9482eb69b08b1702b33a571bfe8525bbdf572a2f8f5feb7f4e29\": rpc error: code = NotFound desc = could not find container \"8b1dd6490cae9482eb69b08b1702b33a571bfe8525bbdf572a2f8f5feb7f4e29\": container with ID starting with 8b1dd6490cae9482eb69b08b1702b33a571bfe8525bbdf572a2f8f5feb7f4e29 not found: ID does not exist" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.404018 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.407782 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495070-b47jt" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.491547 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p55ww\" (UniqueName: \"kubernetes.io/projected/c5a3f592-54c6-44f1-9f09-f366502287a6-kube-api-access-p55ww\") pod \"c5a3f592-54c6-44f1-9f09-f366502287a6\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.491597 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqqvq\" (UniqueName: \"kubernetes.io/projected/76e30b28-5660-4c1a-a31b-626cc3bb6c38-kube-api-access-nqqvq\") pod \"76e30b28-5660-4c1a-a31b-626cc3bb6c38\" (UID: \"76e30b28-5660-4c1a-a31b-626cc3bb6c38\") " Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.491623 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76e30b28-5660-4c1a-a31b-626cc3bb6c38-config-volume\") pod \"76e30b28-5660-4c1a-a31b-626cc3bb6c38\" (UID: \"76e30b28-5660-4c1a-a31b-626cc3bb6c38\") " Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.491709 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c5a3f592-54c6-44f1-9f09-f366502287a6-etc-machine-id\") pod \"c5a3f592-54c6-44f1-9f09-f366502287a6\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.491741 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5a3f592-54c6-44f1-9f09-f366502287a6-config-data-custom\") pod \"c5a3f592-54c6-44f1-9f09-f366502287a6\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.491758 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5a3f592-54c6-44f1-9f09-f366502287a6-config-data\") pod \"c5a3f592-54c6-44f1-9f09-f366502287a6\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.491794 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5a3f592-54c6-44f1-9f09-f366502287a6-logs\") pod \"c5a3f592-54c6-44f1-9f09-f366502287a6\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.491819 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76e30b28-5660-4c1a-a31b-626cc3bb6c38-secret-volume\") pod \"76e30b28-5660-4c1a-a31b-626cc3bb6c38\" (UID: \"76e30b28-5660-4c1a-a31b-626cc3bb6c38\") " Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.491847 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5a3f592-54c6-44f1-9f09-f366502287a6-scripts\") pod \"c5a3f592-54c6-44f1-9f09-f366502287a6\" (UID: \"c5a3f592-54c6-44f1-9f09-f366502287a6\") " Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.492549 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5a3f592-54c6-44f1-9f09-f366502287a6-logs" (OuterVolumeSpecName: "logs") pod "c5a3f592-54c6-44f1-9f09-f366502287a6" (UID: "c5a3f592-54c6-44f1-9f09-f366502287a6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.493142 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76e30b28-5660-4c1a-a31b-626cc3bb6c38-config-volume" (OuterVolumeSpecName: "config-volume") pod "76e30b28-5660-4c1a-a31b-626cc3bb6c38" (UID: "76e30b28-5660-4c1a-a31b-626cc3bb6c38"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.493203 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5a3f592-54c6-44f1-9f09-f366502287a6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c5a3f592-54c6-44f1-9f09-f366502287a6" (UID: "c5a3f592-54c6-44f1-9f09-f366502287a6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.495204 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5a3f592-54c6-44f1-9f09-f366502287a6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c5a3f592-54c6-44f1-9f09-f366502287a6" (UID: "c5a3f592-54c6-44f1-9f09-f366502287a6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.495230 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76e30b28-5660-4c1a-a31b-626cc3bb6c38-kube-api-access-nqqvq" (OuterVolumeSpecName: "kube-api-access-nqqvq") pod "76e30b28-5660-4c1a-a31b-626cc3bb6c38" (UID: "76e30b28-5660-4c1a-a31b-626cc3bb6c38"). InnerVolumeSpecName "kube-api-access-nqqvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.495478 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76e30b28-5660-4c1a-a31b-626cc3bb6c38-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "76e30b28-5660-4c1a-a31b-626cc3bb6c38" (UID: "76e30b28-5660-4c1a-a31b-626cc3bb6c38"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.496099 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5a3f592-54c6-44f1-9f09-f366502287a6-scripts" (OuterVolumeSpecName: "scripts") pod "c5a3f592-54c6-44f1-9f09-f366502287a6" (UID: "c5a3f592-54c6-44f1-9f09-f366502287a6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.496898 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5a3f592-54c6-44f1-9f09-f366502287a6-kube-api-access-p55ww" (OuterVolumeSpecName: "kube-api-access-p55ww") pod "c5a3f592-54c6-44f1-9f09-f366502287a6" (UID: "c5a3f592-54c6-44f1-9f09-f366502287a6"). InnerVolumeSpecName "kube-api-access-p55ww". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.528119 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5a3f592-54c6-44f1-9f09-f366502287a6-config-data" (OuterVolumeSpecName: "config-data") pod "c5a3f592-54c6-44f1-9f09-f366502287a6" (UID: "c5a3f592-54c6-44f1-9f09-f366502287a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.593406 4714 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76e30b28-5660-4c1a-a31b-626cc3bb6c38-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.593442 4714 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5a3f592-54c6-44f1-9f09-f366502287a6-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.593454 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p55ww\" (UniqueName: \"kubernetes.io/projected/c5a3f592-54c6-44f1-9f09-f366502287a6-kube-api-access-p55ww\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.593468 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqqvq\" (UniqueName: \"kubernetes.io/projected/76e30b28-5660-4c1a-a31b-626cc3bb6c38-kube-api-access-nqqvq\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.593478 4714 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76e30b28-5660-4c1a-a31b-626cc3bb6c38-config-volume\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.593486 4714 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c5a3f592-54c6-44f1-9f09-f366502287a6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.593496 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5a3f592-54c6-44f1-9f09-f366502287a6-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.593503 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5a3f592-54c6-44f1-9f09-f366502287a6-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:02 crc kubenswrapper[4714]: I0129 16:30:02.593511 4714 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5a3f592-54c6-44f1-9f09-f366502287a6-logs\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.168264 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29495070-b47jt" event={"ID":"76e30b28-5660-4c1a-a31b-626cc3bb6c38","Type":"ContainerDied","Data":"fcb9acc88fce040f1e751e565b0e7e67527c868c890f2f7bba33c01e544174b1"} Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.168324 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495070-b47jt" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.168347 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcb9acc88fce040f1e751e565b0e7e67527c868c890f2f7bba33c01e544174b1" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.170670 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"c5a3f592-54c6-44f1-9f09-f366502287a6","Type":"ContainerDied","Data":"fcf337a11251f351330506928dad191d61b9373e215ea14ec246e54e5e9a3034"} Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.170752 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.170766 4714 scope.go:117] "RemoveContainer" containerID="922b00d2d16cf09fabe14b15b7a7648c66244d5a615d55c77b2cc333c2095cf3" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.193219 4714 scope.go:117] "RemoveContainer" containerID="32d98c19e9cd977d60b8a6256b941e6d21b55abe95ffb10609d846a79c267c86" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.210068 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.215696 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.852199 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-64s46"] Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.865709 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-64s46"] Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.877069 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder9de3-account-delete-6gpgf"] Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.883802 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-9de3-account-create-update-79889"] Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.889461 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder9de3-account-delete-6gpgf"] Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.895619 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-9de3-account-create-update-79889"] Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.938360 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-db-create-hqbqv"] Jan 29 16:30:03 crc kubenswrapper[4714]: E0129 16:30:03.938709 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d989387-6924-48df-a1d4-1c63911dd476" containerName="mariadb-account-delete" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.938737 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d989387-6924-48df-a1d4-1c63911dd476" containerName="mariadb-account-delete" Jan 29 16:30:03 crc kubenswrapper[4714]: E0129 16:30:03.938766 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0390b29-ac12-4c76-a954-8c7236d81661" containerName="probe" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.938780 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0390b29-ac12-4c76-a954-8c7236d81661" containerName="probe" Jan 29 16:30:03 crc kubenswrapper[4714]: E0129 16:30:03.938798 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e01c83c7-65ba-4f1b-9d17-ba5a824216bb" containerName="probe" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.938809 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="e01c83c7-65ba-4f1b-9d17-ba5a824216bb" containerName="probe" Jan 29 16:30:03 crc kubenswrapper[4714]: E0129 16:30:03.938825 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5a3f592-54c6-44f1-9f09-f366502287a6" containerName="cinder-api-log" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.938836 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5a3f592-54c6-44f1-9f09-f366502287a6" containerName="cinder-api-log" Jan 29 16:30:03 crc kubenswrapper[4714]: E0129 16:30:03.938852 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5a3f592-54c6-44f1-9f09-f366502287a6" containerName="cinder-api" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.938862 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5a3f592-54c6-44f1-9f09-f366502287a6" containerName="cinder-api" Jan 29 16:30:03 crc kubenswrapper[4714]: E0129 16:30:03.938881 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76e30b28-5660-4c1a-a31b-626cc3bb6c38" containerName="collect-profiles" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.938892 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="76e30b28-5660-4c1a-a31b-626cc3bb6c38" containerName="collect-profiles" Jan 29 16:30:03 crc kubenswrapper[4714]: E0129 16:30:03.938909 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e01c83c7-65ba-4f1b-9d17-ba5a824216bb" containerName="cinder-backup" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.938919 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="e01c83c7-65ba-4f1b-9d17-ba5a824216bb" containerName="cinder-backup" Jan 29 16:30:03 crc kubenswrapper[4714]: E0129 16:30:03.938959 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0390b29-ac12-4c76-a954-8c7236d81661" containerName="cinder-scheduler" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.938973 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0390b29-ac12-4c76-a954-8c7236d81661" containerName="cinder-scheduler" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.939169 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="cinder-volume" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.939193 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0390b29-ac12-4c76-a954-8c7236d81661" containerName="probe" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.939213 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="e01c83c7-65ba-4f1b-9d17-ba5a824216bb" containerName="cinder-backup" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.939228 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5a3f592-54c6-44f1-9f09-f366502287a6" containerName="cinder-api" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.939244 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d989387-6924-48df-a1d4-1c63911dd476" containerName="mariadb-account-delete" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.939258 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5a3f592-54c6-44f1-9f09-f366502287a6" containerName="cinder-api-log" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.939272 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="76e30b28-5660-4c1a-a31b-626cc3bb6c38" containerName="collect-profiles" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.939288 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0390b29-ac12-4c76-a954-8c7236d81661" containerName="cinder-scheduler" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.939306 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="e01c83c7-65ba-4f1b-9d17-ba5a824216bb" containerName="probe" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.939324 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e154d80-4b79-4f74-809e-c1c274ed4063" containerName="probe" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.940052 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-hqbqv" Jan 29 16:30:03 crc kubenswrapper[4714]: I0129 16:30:03.944141 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-hqbqv"] Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.037199 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4vmr\" (UniqueName: \"kubernetes.io/projected/5e9c54b1-972a-4807-90af-f94a884002bd-kube-api-access-g4vmr\") pod \"cinder-db-create-hqbqv\" (UID: \"5e9c54b1-972a-4807-90af-f94a884002bd\") " pod="cinder-kuttl-tests/cinder-db-create-hqbqv" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.037268 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e9c54b1-972a-4807-90af-f94a884002bd-operator-scripts\") pod \"cinder-db-create-hqbqv\" (UID: \"5e9c54b1-972a-4807-90af-f94a884002bd\") " pod="cinder-kuttl-tests/cinder-db-create-hqbqv" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.054762 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-6e48-account-create-update-qff9j"] Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.055529 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-6e48-account-create-update-qff9j" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.057861 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-db-secret" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.073607 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-6e48-account-create-update-qff9j"] Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.138813 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3306850-8843-48e1-b203-7f52de72682f-operator-scripts\") pod \"cinder-6e48-account-create-update-qff9j\" (UID: \"c3306850-8843-48e1-b203-7f52de72682f\") " pod="cinder-kuttl-tests/cinder-6e48-account-create-update-qff9j" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.139107 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4vmr\" (UniqueName: \"kubernetes.io/projected/5e9c54b1-972a-4807-90af-f94a884002bd-kube-api-access-g4vmr\") pod \"cinder-db-create-hqbqv\" (UID: \"5e9c54b1-972a-4807-90af-f94a884002bd\") " pod="cinder-kuttl-tests/cinder-db-create-hqbqv" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.139175 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pztc9\" (UniqueName: \"kubernetes.io/projected/c3306850-8843-48e1-b203-7f52de72682f-kube-api-access-pztc9\") pod \"cinder-6e48-account-create-update-qff9j\" (UID: \"c3306850-8843-48e1-b203-7f52de72682f\") " pod="cinder-kuttl-tests/cinder-6e48-account-create-update-qff9j" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.139237 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e9c54b1-972a-4807-90af-f94a884002bd-operator-scripts\") pod \"cinder-db-create-hqbqv\" (UID: \"5e9c54b1-972a-4807-90af-f94a884002bd\") " pod="cinder-kuttl-tests/cinder-db-create-hqbqv" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.140184 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e9c54b1-972a-4807-90af-f94a884002bd-operator-scripts\") pod \"cinder-db-create-hqbqv\" (UID: \"5e9c54b1-972a-4807-90af-f94a884002bd\") " pod="cinder-kuttl-tests/cinder-db-create-hqbqv" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.162647 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4vmr\" (UniqueName: \"kubernetes.io/projected/5e9c54b1-972a-4807-90af-f94a884002bd-kube-api-access-g4vmr\") pod \"cinder-db-create-hqbqv\" (UID: \"5e9c54b1-972a-4807-90af-f94a884002bd\") " pod="cinder-kuttl-tests/cinder-db-create-hqbqv" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.193417 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d989387-6924-48df-a1d4-1c63911dd476" path="/var/lib/kubelet/pods/8d989387-6924-48df-a1d4-1c63911dd476/volumes" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.194331 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0390b29-ac12-4c76-a954-8c7236d81661" path="/var/lib/kubelet/pods/c0390b29-ac12-4c76-a954-8c7236d81661/volumes" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.195135 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5a3f592-54c6-44f1-9f09-f366502287a6" path="/var/lib/kubelet/pods/c5a3f592-54c6-44f1-9f09-f366502287a6/volumes" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.196257 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c747693b-f2e9-4073-9432-115643a6b6d1" path="/var/lib/kubelet/pods/c747693b-f2e9-4073-9432-115643a6b6d1/volumes" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.196749 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e670557b-650e-478c-9f87-eaba6641f02f" path="/var/lib/kubelet/pods/e670557b-650e-478c-9f87-eaba6641f02f/volumes" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.240917 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pztc9\" (UniqueName: \"kubernetes.io/projected/c3306850-8843-48e1-b203-7f52de72682f-kube-api-access-pztc9\") pod \"cinder-6e48-account-create-update-qff9j\" (UID: \"c3306850-8843-48e1-b203-7f52de72682f\") " pod="cinder-kuttl-tests/cinder-6e48-account-create-update-qff9j" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.241047 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3306850-8843-48e1-b203-7f52de72682f-operator-scripts\") pod \"cinder-6e48-account-create-update-qff9j\" (UID: \"c3306850-8843-48e1-b203-7f52de72682f\") " pod="cinder-kuttl-tests/cinder-6e48-account-create-update-qff9j" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.241834 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3306850-8843-48e1-b203-7f52de72682f-operator-scripts\") pod \"cinder-6e48-account-create-update-qff9j\" (UID: \"c3306850-8843-48e1-b203-7f52de72682f\") " pod="cinder-kuttl-tests/cinder-6e48-account-create-update-qff9j" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.258122 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pztc9\" (UniqueName: \"kubernetes.io/projected/c3306850-8843-48e1-b203-7f52de72682f-kube-api-access-pztc9\") pod \"cinder-6e48-account-create-update-qff9j\" (UID: \"c3306850-8843-48e1-b203-7f52de72682f\") " pod="cinder-kuttl-tests/cinder-6e48-account-create-update-qff9j" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.264981 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-hqbqv" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.375492 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-6e48-account-create-update-qff9j" Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.600898 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-6e48-account-create-update-qff9j"] Jan 29 16:30:04 crc kubenswrapper[4714]: W0129 16:30:04.604251 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3306850_8843_48e1_b203_7f52de72682f.slice/crio-802cb29550e90931d915fdf150b111a01142b90181c1330445c631bbf924410b WatchSource:0}: Error finding container 802cb29550e90931d915fdf150b111a01142b90181c1330445c631bbf924410b: Status 404 returned error can't find the container with id 802cb29550e90931d915fdf150b111a01142b90181c1330445c631bbf924410b Jan 29 16:30:04 crc kubenswrapper[4714]: I0129 16:30:04.694413 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-hqbqv"] Jan 29 16:30:04 crc kubenswrapper[4714]: W0129 16:30:04.695873 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e9c54b1_972a_4807_90af_f94a884002bd.slice/crio-cd3ed4ce144cba24bd2fd48ece0ee79d8f7dc5583431a0cba4fe696107cd50a3 WatchSource:0}: Error finding container cd3ed4ce144cba24bd2fd48ece0ee79d8f7dc5583431a0cba4fe696107cd50a3: Status 404 returned error can't find the container with id cd3ed4ce144cba24bd2fd48ece0ee79d8f7dc5583431a0cba4fe696107cd50a3 Jan 29 16:30:05 crc kubenswrapper[4714]: I0129 16:30:05.187777 4714 generic.go:334] "Generic (PLEG): container finished" podID="c3306850-8843-48e1-b203-7f52de72682f" containerID="b09b52cf99e966280d15ceb6a6529b45a9303070f260a1e02acc4b1cf0da02c3" exitCode=0 Jan 29 16:30:05 crc kubenswrapper[4714]: I0129 16:30:05.187860 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-6e48-account-create-update-qff9j" event={"ID":"c3306850-8843-48e1-b203-7f52de72682f","Type":"ContainerDied","Data":"b09b52cf99e966280d15ceb6a6529b45a9303070f260a1e02acc4b1cf0da02c3"} Jan 29 16:30:05 crc kubenswrapper[4714]: I0129 16:30:05.187892 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-6e48-account-create-update-qff9j" event={"ID":"c3306850-8843-48e1-b203-7f52de72682f","Type":"ContainerStarted","Data":"802cb29550e90931d915fdf150b111a01142b90181c1330445c631bbf924410b"} Jan 29 16:30:05 crc kubenswrapper[4714]: I0129 16:30:05.189536 4714 generic.go:334] "Generic (PLEG): container finished" podID="5e9c54b1-972a-4807-90af-f94a884002bd" containerID="ad6e6492e17aa0045196d2d7816583e0511c88fa1ba9566c638560f377a604b8" exitCode=0 Jan 29 16:30:05 crc kubenswrapper[4714]: I0129 16:30:05.189564 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-hqbqv" event={"ID":"5e9c54b1-972a-4807-90af-f94a884002bd","Type":"ContainerDied","Data":"ad6e6492e17aa0045196d2d7816583e0511c88fa1ba9566c638560f377a604b8"} Jan 29 16:30:05 crc kubenswrapper[4714]: I0129 16:30:05.189580 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-hqbqv" event={"ID":"5e9c54b1-972a-4807-90af-f94a884002bd","Type":"ContainerStarted","Data":"cd3ed4ce144cba24bd2fd48ece0ee79d8f7dc5583431a0cba4fe696107cd50a3"} Jan 29 16:30:06 crc kubenswrapper[4714]: I0129 16:30:06.518115 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-hqbqv" Jan 29 16:30:06 crc kubenswrapper[4714]: I0129 16:30:06.524019 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-6e48-account-create-update-qff9j" Jan 29 16:30:06 crc kubenswrapper[4714]: I0129 16:30:06.676879 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pztc9\" (UniqueName: \"kubernetes.io/projected/c3306850-8843-48e1-b203-7f52de72682f-kube-api-access-pztc9\") pod \"c3306850-8843-48e1-b203-7f52de72682f\" (UID: \"c3306850-8843-48e1-b203-7f52de72682f\") " Jan 29 16:30:06 crc kubenswrapper[4714]: I0129 16:30:06.677012 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4vmr\" (UniqueName: \"kubernetes.io/projected/5e9c54b1-972a-4807-90af-f94a884002bd-kube-api-access-g4vmr\") pod \"5e9c54b1-972a-4807-90af-f94a884002bd\" (UID: \"5e9c54b1-972a-4807-90af-f94a884002bd\") " Jan 29 16:30:06 crc kubenswrapper[4714]: I0129 16:30:06.677082 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3306850-8843-48e1-b203-7f52de72682f-operator-scripts\") pod \"c3306850-8843-48e1-b203-7f52de72682f\" (UID: \"c3306850-8843-48e1-b203-7f52de72682f\") " Jan 29 16:30:06 crc kubenswrapper[4714]: I0129 16:30:06.677126 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e9c54b1-972a-4807-90af-f94a884002bd-operator-scripts\") pod \"5e9c54b1-972a-4807-90af-f94a884002bd\" (UID: \"5e9c54b1-972a-4807-90af-f94a884002bd\") " Jan 29 16:30:06 crc kubenswrapper[4714]: I0129 16:30:06.678804 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3306850-8843-48e1-b203-7f52de72682f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c3306850-8843-48e1-b203-7f52de72682f" (UID: "c3306850-8843-48e1-b203-7f52de72682f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:30:06 crc kubenswrapper[4714]: I0129 16:30:06.678873 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e9c54b1-972a-4807-90af-f94a884002bd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5e9c54b1-972a-4807-90af-f94a884002bd" (UID: "5e9c54b1-972a-4807-90af-f94a884002bd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:30:06 crc kubenswrapper[4714]: I0129 16:30:06.685930 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e9c54b1-972a-4807-90af-f94a884002bd-kube-api-access-g4vmr" (OuterVolumeSpecName: "kube-api-access-g4vmr") pod "5e9c54b1-972a-4807-90af-f94a884002bd" (UID: "5e9c54b1-972a-4807-90af-f94a884002bd"). InnerVolumeSpecName "kube-api-access-g4vmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:06 crc kubenswrapper[4714]: I0129 16:30:06.686266 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3306850-8843-48e1-b203-7f52de72682f-kube-api-access-pztc9" (OuterVolumeSpecName: "kube-api-access-pztc9") pod "c3306850-8843-48e1-b203-7f52de72682f" (UID: "c3306850-8843-48e1-b203-7f52de72682f"). InnerVolumeSpecName "kube-api-access-pztc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:06 crc kubenswrapper[4714]: I0129 16:30:06.778871 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pztc9\" (UniqueName: \"kubernetes.io/projected/c3306850-8843-48e1-b203-7f52de72682f-kube-api-access-pztc9\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:06 crc kubenswrapper[4714]: I0129 16:30:06.778915 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4vmr\" (UniqueName: \"kubernetes.io/projected/5e9c54b1-972a-4807-90af-f94a884002bd-kube-api-access-g4vmr\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:06 crc kubenswrapper[4714]: I0129 16:30:06.778953 4714 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3306850-8843-48e1-b203-7f52de72682f-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:06 crc kubenswrapper[4714]: I0129 16:30:06.778972 4714 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e9c54b1-972a-4807-90af-f94a884002bd-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:07 crc kubenswrapper[4714]: I0129 16:30:07.210436 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-hqbqv" event={"ID":"5e9c54b1-972a-4807-90af-f94a884002bd","Type":"ContainerDied","Data":"cd3ed4ce144cba24bd2fd48ece0ee79d8f7dc5583431a0cba4fe696107cd50a3"} Jan 29 16:30:07 crc kubenswrapper[4714]: I0129 16:30:07.210981 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd3ed4ce144cba24bd2fd48ece0ee79d8f7dc5583431a0cba4fe696107cd50a3" Jan 29 16:30:07 crc kubenswrapper[4714]: I0129 16:30:07.211101 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-hqbqv" Jan 29 16:30:07 crc kubenswrapper[4714]: I0129 16:30:07.213389 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-6e48-account-create-update-qff9j" event={"ID":"c3306850-8843-48e1-b203-7f52de72682f","Type":"ContainerDied","Data":"802cb29550e90931d915fdf150b111a01142b90181c1330445c631bbf924410b"} Jan 29 16:30:07 crc kubenswrapper[4714]: I0129 16:30:07.213445 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="802cb29550e90931d915fdf150b111a01142b90181c1330445c631bbf924410b" Jan 29 16:30:07 crc kubenswrapper[4714]: I0129 16:30:07.213467 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-6e48-account-create-update-qff9j" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.295428 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-zpbtx"] Jan 29 16:30:09 crc kubenswrapper[4714]: E0129 16:30:09.295726 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e9c54b1-972a-4807-90af-f94a884002bd" containerName="mariadb-database-create" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.295738 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e9c54b1-972a-4807-90af-f94a884002bd" containerName="mariadb-database-create" Jan 29 16:30:09 crc kubenswrapper[4714]: E0129 16:30:09.295752 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3306850-8843-48e1-b203-7f52de72682f" containerName="mariadb-account-create-update" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.295759 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3306850-8843-48e1-b203-7f52de72682f" containerName="mariadb-account-create-update" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.295884 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e9c54b1-972a-4807-90af-f94a884002bd" containerName="mariadb-database-create" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.296002 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3306850-8843-48e1-b203-7f52de72682f" containerName="mariadb-account-create-update" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.296412 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.298820 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scripts" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.299543 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-config-data" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.299657 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-cinder-dockercfg-thgfp" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.301272 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"combined-ca-bundle" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.317906 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-zpbtx"] Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.416903 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0cabdde2-0578-405a-9147-efe4d1db7e90-etc-machine-id\") pod \"cinder-db-sync-zpbtx\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.417521 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-combined-ca-bundle\") pod \"cinder-db-sync-zpbtx\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.417568 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-db-sync-config-data\") pod \"cinder-db-sync-zpbtx\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.417783 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-scripts\") pod \"cinder-db-sync-zpbtx\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.417904 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5p98\" (UniqueName: \"kubernetes.io/projected/0cabdde2-0578-405a-9147-efe4d1db7e90-kube-api-access-w5p98\") pod \"cinder-db-sync-zpbtx\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.418063 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-config-data\") pod \"cinder-db-sync-zpbtx\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.519738 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-combined-ca-bundle\") pod \"cinder-db-sync-zpbtx\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.519802 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-db-sync-config-data\") pod \"cinder-db-sync-zpbtx\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.519830 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-scripts\") pod \"cinder-db-sync-zpbtx\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.519856 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5p98\" (UniqueName: \"kubernetes.io/projected/0cabdde2-0578-405a-9147-efe4d1db7e90-kube-api-access-w5p98\") pod \"cinder-db-sync-zpbtx\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.519890 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-config-data\") pod \"cinder-db-sync-zpbtx\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.519945 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0cabdde2-0578-405a-9147-efe4d1db7e90-etc-machine-id\") pod \"cinder-db-sync-zpbtx\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.520008 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0cabdde2-0578-405a-9147-efe4d1db7e90-etc-machine-id\") pod \"cinder-db-sync-zpbtx\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.527736 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-scripts\") pod \"cinder-db-sync-zpbtx\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.527781 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-db-sync-config-data\") pod \"cinder-db-sync-zpbtx\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.527870 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-combined-ca-bundle\") pod \"cinder-db-sync-zpbtx\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.528564 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-config-data\") pod \"cinder-db-sync-zpbtx\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.538811 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5p98\" (UniqueName: \"kubernetes.io/projected/0cabdde2-0578-405a-9147-efe4d1db7e90-kube-api-access-w5p98\") pod \"cinder-db-sync-zpbtx\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:09 crc kubenswrapper[4714]: I0129 16:30:09.642799 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:10 crc kubenswrapper[4714]: I0129 16:30:10.065432 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-zpbtx"] Jan 29 16:30:10 crc kubenswrapper[4714]: I0129 16:30:10.240469 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" event={"ID":"0cabdde2-0578-405a-9147-efe4d1db7e90","Type":"ContainerStarted","Data":"546b72ac34d10addd067108b73bef9b904aea8a895447843e1314af1b2a18e00"} Jan 29 16:30:11 crc kubenswrapper[4714]: I0129 16:30:11.252363 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" event={"ID":"0cabdde2-0578-405a-9147-efe4d1db7e90","Type":"ContainerStarted","Data":"be7a968b80d5f3fb2bec436bd6006753f8b129fcc60e1b9be5f43a75c59f2e55"} Jan 29 16:30:11 crc kubenswrapper[4714]: I0129 16:30:11.270632 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" podStartSLOduration=2.2706143770000002 podStartE2EDuration="2.270614377s" podCreationTimestamp="2026-01-29 16:30:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:30:11.268691637 +0000 UTC m=+1217.789192817" watchObservedRunningTime="2026-01-29 16:30:11.270614377 +0000 UTC m=+1217.791115507" Jan 29 16:30:12 crc kubenswrapper[4714]: E0129 16:30:12.188315 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:30:13 crc kubenswrapper[4714]: I0129 16:30:13.269836 4714 generic.go:334] "Generic (PLEG): container finished" podID="0cabdde2-0578-405a-9147-efe4d1db7e90" containerID="be7a968b80d5f3fb2bec436bd6006753f8b129fcc60e1b9be5f43a75c59f2e55" exitCode=0 Jan 29 16:30:13 crc kubenswrapper[4714]: I0129 16:30:13.269873 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" event={"ID":"0cabdde2-0578-405a-9147-efe4d1db7e90","Type":"ContainerDied","Data":"be7a968b80d5f3fb2bec436bd6006753f8b129fcc60e1b9be5f43a75c59f2e55"} Jan 29 16:30:14 crc kubenswrapper[4714]: I0129 16:30:14.623246 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:14 crc kubenswrapper[4714]: I0129 16:30:14.715515 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0cabdde2-0578-405a-9147-efe4d1db7e90-etc-machine-id\") pod \"0cabdde2-0578-405a-9147-efe4d1db7e90\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " Jan 29 16:30:14 crc kubenswrapper[4714]: I0129 16:30:14.715637 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cabdde2-0578-405a-9147-efe4d1db7e90-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0cabdde2-0578-405a-9147-efe4d1db7e90" (UID: "0cabdde2-0578-405a-9147-efe4d1db7e90"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:14 crc kubenswrapper[4714]: I0129 16:30:14.715677 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-db-sync-config-data\") pod \"0cabdde2-0578-405a-9147-efe4d1db7e90\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " Jan 29 16:30:14 crc kubenswrapper[4714]: I0129 16:30:14.715834 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-combined-ca-bundle\") pod \"0cabdde2-0578-405a-9147-efe4d1db7e90\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " Jan 29 16:30:14 crc kubenswrapper[4714]: I0129 16:30:14.715859 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5p98\" (UniqueName: \"kubernetes.io/projected/0cabdde2-0578-405a-9147-efe4d1db7e90-kube-api-access-w5p98\") pod \"0cabdde2-0578-405a-9147-efe4d1db7e90\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " Jan 29 16:30:14 crc kubenswrapper[4714]: I0129 16:30:14.715908 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-scripts\") pod \"0cabdde2-0578-405a-9147-efe4d1db7e90\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " Jan 29 16:30:14 crc kubenswrapper[4714]: I0129 16:30:14.715975 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-config-data\") pod \"0cabdde2-0578-405a-9147-efe4d1db7e90\" (UID: \"0cabdde2-0578-405a-9147-efe4d1db7e90\") " Jan 29 16:30:14 crc kubenswrapper[4714]: I0129 16:30:14.716384 4714 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0cabdde2-0578-405a-9147-efe4d1db7e90-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:14 crc kubenswrapper[4714]: I0129 16:30:14.720847 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cabdde2-0578-405a-9147-efe4d1db7e90-kube-api-access-w5p98" (OuterVolumeSpecName: "kube-api-access-w5p98") pod "0cabdde2-0578-405a-9147-efe4d1db7e90" (UID: "0cabdde2-0578-405a-9147-efe4d1db7e90"). InnerVolumeSpecName "kube-api-access-w5p98". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:14 crc kubenswrapper[4714]: I0129 16:30:14.720860 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0cabdde2-0578-405a-9147-efe4d1db7e90" (UID: "0cabdde2-0578-405a-9147-efe4d1db7e90"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:14 crc kubenswrapper[4714]: I0129 16:30:14.721018 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-scripts" (OuterVolumeSpecName: "scripts") pod "0cabdde2-0578-405a-9147-efe4d1db7e90" (UID: "0cabdde2-0578-405a-9147-efe4d1db7e90"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:14 crc kubenswrapper[4714]: I0129 16:30:14.734254 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0cabdde2-0578-405a-9147-efe4d1db7e90" (UID: "0cabdde2-0578-405a-9147-efe4d1db7e90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:14 crc kubenswrapper[4714]: I0129 16:30:14.747695 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-config-data" (OuterVolumeSpecName: "config-data") pod "0cabdde2-0578-405a-9147-efe4d1db7e90" (UID: "0cabdde2-0578-405a-9147-efe4d1db7e90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:14 crc kubenswrapper[4714]: I0129 16:30:14.817262 4714 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:14 crc kubenswrapper[4714]: I0129 16:30:14.817300 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:14 crc kubenswrapper[4714]: I0129 16:30:14.817313 4714 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:14 crc kubenswrapper[4714]: I0129 16:30:14.817323 4714 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cabdde2-0578-405a-9147-efe4d1db7e90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:14 crc kubenswrapper[4714]: I0129 16:30:14.817332 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5p98\" (UniqueName: \"kubernetes.io/projected/0cabdde2-0578-405a-9147-efe4d1db7e90-kube-api-access-w5p98\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.288458 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" event={"ID":"0cabdde2-0578-405a-9147-efe4d1db7e90","Type":"ContainerDied","Data":"546b72ac34d10addd067108b73bef9b904aea8a895447843e1314af1b2a18e00"} Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.288499 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="546b72ac34d10addd067108b73bef9b904aea8a895447843e1314af1b2a18e00" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.288557 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-zpbtx" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.539075 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Jan 29 16:30:15 crc kubenswrapper[4714]: E0129 16:30:15.539764 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cabdde2-0578-405a-9147-efe4d1db7e90" containerName="cinder-db-sync" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.539793 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cabdde2-0578-405a-9147-efe4d1db7e90" containerName="cinder-db-sync" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.540032 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cabdde2-0578-405a-9147-efe4d1db7e90" containerName="cinder-db-sync" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.540892 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.543627 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-config-data" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.551816 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"combined-ca-bundle" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.552055 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scripts" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.552285 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-cinder-dockercfg-thgfp" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.556272 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scheduler-config-data" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.570670 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.581405 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.594840 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.596531 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-backup-config-data" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.625076 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.630027 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-scripts\") pod \"cinder-scheduler-0\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.630321 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/130b46c6-e7e5-4202-bea4-1214ec4766e8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.630462 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-config-data\") pod \"cinder-scheduler-0\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.630605 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.630727 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85zz9\" (UniqueName: \"kubernetes.io/projected/130b46c6-e7e5-4202-bea4-1214ec4766e8-kube-api-access-85zz9\") pod \"cinder-scheduler-0\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.630843 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.647977 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.649029 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.653796 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-volume-volume1-config-data" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.663422 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732201 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732250 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732274 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732297 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-run\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732321 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732341 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-run\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732372 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-config-data\") pod \"cinder-scheduler-0\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732402 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732421 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-dev\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732437 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732457 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-lib-modules\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732487 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-dev\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732506 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732527 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732555 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732575 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732594 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732617 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-config-data-custom\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732644 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-scripts\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732668 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732695 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl9k4\" (UniqueName: \"kubernetes.io/projected/7c2fde86-e7c8-4605-9750-8464ca4b7d58-kube-api-access-bl9k4\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732720 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732745 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85zz9\" (UniqueName: \"kubernetes.io/projected/130b46c6-e7e5-4202-bea4-1214ec4766e8-kube-api-access-85zz9\") pod \"cinder-scheduler-0\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732769 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732793 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732822 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732847 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732878 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732915 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-etc-nvme\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732959 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-scripts\") pod \"cinder-scheduler-0\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.732993 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/130b46c6-e7e5-4202-bea4-1214ec4766e8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.733015 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-config-data\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.733048 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.733070 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl96t\" (UniqueName: \"kubernetes.io/projected/27df3c21-5ecb-4af0-9e48-a40f826dc75d-kube-api-access-jl96t\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.733089 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-sys\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.733111 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-sys\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.733964 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/130b46c6-e7e5-4202-bea4-1214ec4766e8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.742725 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.743206 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.744724 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-scripts\") pod \"cinder-scheduler-0\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.748814 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-config-data\") pod \"cinder-scheduler-0\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.754002 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.755408 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.758853 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-api-config-data" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.759050 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cert-cinder-public-svc" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.759855 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cert-cinder-internal-svc" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.762335 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.769812 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85zz9\" (UniqueName: \"kubernetes.io/projected/130b46c6-e7e5-4202-bea4-1214ec4766e8-kube-api-access-85zz9\") pod \"cinder-scheduler-0\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.834286 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-config-data\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.834351 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.834375 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl96t\" (UniqueName: \"kubernetes.io/projected/27df3c21-5ecb-4af0-9e48-a40f826dc75d-kube-api-access-jl96t\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.834395 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-sys\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.834416 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-sys\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.834435 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.834456 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.834477 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.834501 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/93eb21c7-d0f9-4648-a671-03d3ccd28429-etc-machine-id\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.834519 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-config-data\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835037 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-config-data-custom\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835124 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-run\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835151 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-public-tls-certs\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835171 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835189 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835246 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-run\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835297 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835319 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-dev\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835344 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835365 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-lib-modules\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835399 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-dev\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835454 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835481 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835504 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835550 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835577 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835608 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835642 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-config-data-custom\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835682 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-scripts\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835716 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-scripts\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835753 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl9k4\" (UniqueName: \"kubernetes.io/projected/7c2fde86-e7c8-4605-9750-8464ca4b7d58-kube-api-access-bl9k4\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835790 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835829 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835854 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835884 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93eb21c7-d0f9-4648-a671-03d3ccd28429-logs\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.835927 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-str9z\" (UniqueName: \"kubernetes.io/projected/93eb21c7-d0f9-4648-a671-03d3ccd28429-kube-api-access-str9z\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.836008 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.836033 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.836071 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.836109 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-etc-nvme\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.836259 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-etc-nvme\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.836307 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-run\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.836340 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.836366 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-run\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.836411 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.836441 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-dev\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.836469 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.836500 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-lib-modules\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.836526 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-dev\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.836570 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.836597 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.836628 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.836695 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.836728 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-sys\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.836744 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.837201 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.837581 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.837723 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.837766 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-sys\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.838697 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-config-data\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.840140 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.840462 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-scripts\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.840481 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.841146 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-config-data-custom\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.841543 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.843740 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.848541 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.855849 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl96t\" (UniqueName: \"kubernetes.io/projected/27df3c21-5ecb-4af0-9e48-a40f826dc75d-kube-api-access-jl96t\") pod \"cinder-volume-volume1-0\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.859530 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl9k4\" (UniqueName: \"kubernetes.io/projected/7c2fde86-e7c8-4605-9750-8464ca4b7d58-kube-api-access-bl9k4\") pod \"cinder-backup-0\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.861179 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.922711 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.937914 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/93eb21c7-d0f9-4648-a671-03d3ccd28429-etc-machine-id\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.937993 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-config-data\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.938016 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-config-data-custom\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.938022 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/93eb21c7-d0f9-4648-a671-03d3ccd28429-etc-machine-id\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.938046 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-public-tls-certs\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.938154 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.938231 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-scripts\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.938289 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93eb21c7-d0f9-4648-a671-03d3ccd28429-logs\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.938326 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-str9z\" (UniqueName: \"kubernetes.io/projected/93eb21c7-d0f9-4648-a671-03d3ccd28429-kube-api-access-str9z\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.938358 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.939247 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93eb21c7-d0f9-4648-a671-03d3ccd28429-logs\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.945201 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.945216 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-public-tls-certs\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.946374 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-config-data\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.946654 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.949242 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-scripts\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.956632 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-config-data-custom\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.961673 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:15 crc kubenswrapper[4714]: I0129 16:30:15.976619 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-str9z\" (UniqueName: \"kubernetes.io/projected/93eb21c7-d0f9-4648-a671-03d3ccd28429-kube-api-access-str9z\") pod \"cinder-api-0\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:16 crc kubenswrapper[4714]: I0129 16:30:16.125315 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:16 crc kubenswrapper[4714]: I0129 16:30:16.319172 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Jan 29 16:30:16 crc kubenswrapper[4714]: I0129 16:30:16.418206 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Jan 29 16:30:16 crc kubenswrapper[4714]: W0129 16:30:16.428473 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c2fde86_e7c8_4605_9750_8464ca4b7d58.slice/crio-e25204d76e2148fee59f9d4ff12c3c8594f24cfcc42364e257052da501eeec8a WatchSource:0}: Error finding container e25204d76e2148fee59f9d4ff12c3c8594f24cfcc42364e257052da501eeec8a: Status 404 returned error can't find the container with id e25204d76e2148fee59f9d4ff12c3c8594f24cfcc42364e257052da501eeec8a Jan 29 16:30:16 crc kubenswrapper[4714]: I0129 16:30:16.526945 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Jan 29 16:30:16 crc kubenswrapper[4714]: W0129 16:30:16.536012 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27df3c21_5ecb_4af0_9e48_a40f826dc75d.slice/crio-a215b87a9f7bec152d91f39a039625249d594cfc1ffc1ed4fb03f005b6ae19e1 WatchSource:0}: Error finding container a215b87a9f7bec152d91f39a039625249d594cfc1ffc1ed4fb03f005b6ae19e1: Status 404 returned error can't find the container with id a215b87a9f7bec152d91f39a039625249d594cfc1ffc1ed4fb03f005b6ae19e1 Jan 29 16:30:16 crc kubenswrapper[4714]: I0129 16:30:16.660463 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Jan 29 16:30:16 crc kubenswrapper[4714]: W0129 16:30:16.668809 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93eb21c7_d0f9_4648_a671_03d3ccd28429.slice/crio-9cf9a3f95739dab5cf604ede39f2741a3a785d75007c516172ae75ff6b3665b1 WatchSource:0}: Error finding container 9cf9a3f95739dab5cf604ede39f2741a3a785d75007c516172ae75ff6b3665b1: Status 404 returned error can't find the container with id 9cf9a3f95739dab5cf604ede39f2741a3a785d75007c516172ae75ff6b3665b1 Jan 29 16:30:17 crc kubenswrapper[4714]: I0129 16:30:17.305056 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"27df3c21-5ecb-4af0-9e48-a40f826dc75d","Type":"ContainerStarted","Data":"4565e725bbc777847d5f682f8f2a88a1faea9eac4df83a536e5f10e10af89617"} Jan 29 16:30:17 crc kubenswrapper[4714]: I0129 16:30:17.305597 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"27df3c21-5ecb-4af0-9e48-a40f826dc75d","Type":"ContainerStarted","Data":"6825612e424e7d18534f6766e6735d779be6e5909c3365b087bf62b9bfd4f305"} Jan 29 16:30:17 crc kubenswrapper[4714]: I0129 16:30:17.305610 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"27df3c21-5ecb-4af0-9e48-a40f826dc75d","Type":"ContainerStarted","Data":"a215b87a9f7bec152d91f39a039625249d594cfc1ffc1ed4fb03f005b6ae19e1"} Jan 29 16:30:17 crc kubenswrapper[4714]: I0129 16:30:17.312502 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"93eb21c7-d0f9-4648-a671-03d3ccd28429","Type":"ContainerStarted","Data":"b54fd3f2968e79a20a0b8cb352bdc32d69a2daf77ebdec8342b8fb48a253ab80"} Jan 29 16:30:17 crc kubenswrapper[4714]: I0129 16:30:17.312551 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"93eb21c7-d0f9-4648-a671-03d3ccd28429","Type":"ContainerStarted","Data":"9cf9a3f95739dab5cf604ede39f2741a3a785d75007c516172ae75ff6b3665b1"} Jan 29 16:30:17 crc kubenswrapper[4714]: I0129 16:30:17.333352 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"7c2fde86-e7c8-4605-9750-8464ca4b7d58","Type":"ContainerStarted","Data":"c236db1e5dbb8d372959347da33330e9e06dab48278781aa11dc0c6a6f372af8"} Jan 29 16:30:17 crc kubenswrapper[4714]: I0129 16:30:17.333409 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"7c2fde86-e7c8-4605-9750-8464ca4b7d58","Type":"ContainerStarted","Data":"2caedfb03564966ae2d6b87961e4e7d74fd983fe0c16c41eb96e2d6ca6ebd267"} Jan 29 16:30:17 crc kubenswrapper[4714]: I0129 16:30:17.333425 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"7c2fde86-e7c8-4605-9750-8464ca4b7d58","Type":"ContainerStarted","Data":"e25204d76e2148fee59f9d4ff12c3c8594f24cfcc42364e257052da501eeec8a"} Jan 29 16:30:17 crc kubenswrapper[4714]: I0129 16:30:17.335369 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podStartSLOduration=2.335355013 podStartE2EDuration="2.335355013s" podCreationTimestamp="2026-01-29 16:30:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:30:17.324560676 +0000 UTC m=+1223.845061796" watchObservedRunningTime="2026-01-29 16:30:17.335355013 +0000 UTC m=+1223.855856133" Jan 29 16:30:17 crc kubenswrapper[4714]: I0129 16:30:17.343085 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"130b46c6-e7e5-4202-bea4-1214ec4766e8","Type":"ContainerStarted","Data":"bccd615573c7599b167665a1108953cb3d1759c6fb3d978bae9a4b7e75fbc11a"} Jan 29 16:30:17 crc kubenswrapper[4714]: I0129 16:30:17.343125 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"130b46c6-e7e5-4202-bea4-1214ec4766e8","Type":"ContainerStarted","Data":"4117860f9b4009b5187f645af44ec4d5a39f7166f668190572db33d3a8f8e24c"} Jan 29 16:30:17 crc kubenswrapper[4714]: I0129 16:30:17.366715 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-backup-0" podStartSLOduration=2.3666976330000002 podStartE2EDuration="2.366697633s" podCreationTimestamp="2026-01-29 16:30:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:30:17.359308807 +0000 UTC m=+1223.879809937" watchObservedRunningTime="2026-01-29 16:30:17.366697633 +0000 UTC m=+1223.887198753" Jan 29 16:30:18 crc kubenswrapper[4714]: I0129 16:30:18.373614 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"130b46c6-e7e5-4202-bea4-1214ec4766e8","Type":"ContainerStarted","Data":"bbf7723cee6103ad631b2d89beaf6a72a626280c58f32453fc92b4acddbf7202"} Jan 29 16:30:18 crc kubenswrapper[4714]: I0129 16:30:18.378728 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"93eb21c7-d0f9-4648-a671-03d3ccd28429","Type":"ContainerStarted","Data":"0c9e8f216ea8d08fef702164fee1c489d59f45d35804acfb1d8d5c8a08af65d1"} Jan 29 16:30:18 crc kubenswrapper[4714]: I0129 16:30:18.398154 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-scheduler-0" podStartSLOduration=3.39813713 podStartE2EDuration="3.39813713s" podCreationTimestamp="2026-01-29 16:30:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:30:18.396986616 +0000 UTC m=+1224.917487736" watchObservedRunningTime="2026-01-29 16:30:18.39813713 +0000 UTC m=+1224.918638270" Jan 29 16:30:18 crc kubenswrapper[4714]: I0129 16:30:18.424955 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-0" podStartSLOduration=3.424926214 podStartE2EDuration="3.424926214s" podCreationTimestamp="2026-01-29 16:30:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:30:18.42046594 +0000 UTC m=+1224.940967070" watchObservedRunningTime="2026-01-29 16:30:18.424926214 +0000 UTC m=+1224.945427334" Jan 29 16:30:19 crc kubenswrapper[4714]: I0129 16:30:19.387700 4714 generic.go:334] "Generic (PLEG): container finished" podID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" containerID="4565e725bbc777847d5f682f8f2a88a1faea9eac4df83a536e5f10e10af89617" exitCode=1 Jan 29 16:30:19 crc kubenswrapper[4714]: I0129 16:30:19.387740 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"27df3c21-5ecb-4af0-9e48-a40f826dc75d","Type":"ContainerDied","Data":"4565e725bbc777847d5f682f8f2a88a1faea9eac4df83a536e5f10e10af89617"} Jan 29 16:30:19 crc kubenswrapper[4714]: I0129 16:30:19.388339 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:19 crc kubenswrapper[4714]: I0129 16:30:19.388698 4714 scope.go:117] "RemoveContainer" containerID="4565e725bbc777847d5f682f8f2a88a1faea9eac4df83a536e5f10e10af89617" Jan 29 16:30:20 crc kubenswrapper[4714]: I0129 16:30:20.397684 4714 generic.go:334] "Generic (PLEG): container finished" podID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" containerID="6825612e424e7d18534f6766e6735d779be6e5909c3365b087bf62b9bfd4f305" exitCode=1 Jan 29 16:30:20 crc kubenswrapper[4714]: I0129 16:30:20.397885 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"27df3c21-5ecb-4af0-9e48-a40f826dc75d","Type":"ContainerDied","Data":"6825612e424e7d18534f6766e6735d779be6e5909c3365b087bf62b9bfd4f305"} Jan 29 16:30:20 crc kubenswrapper[4714]: I0129 16:30:20.398534 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"27df3c21-5ecb-4af0-9e48-a40f826dc75d","Type":"ContainerStarted","Data":"998fc5de34fad62db822ad95298396d20ef6c7ee52c10eeb679ec05818c2f0ea"} Jan 29 16:30:20 crc kubenswrapper[4714]: I0129 16:30:20.399090 4714 scope.go:117] "RemoveContainer" containerID="6825612e424e7d18534f6766e6735d779be6e5909c3365b087bf62b9bfd4f305" Jan 29 16:30:20 crc kubenswrapper[4714]: I0129 16:30:20.862600 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:20 crc kubenswrapper[4714]: I0129 16:30:20.924031 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:20 crc kubenswrapper[4714]: I0129 16:30:20.963064 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:20 crc kubenswrapper[4714]: I0129 16:30:20.963139 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:21 crc kubenswrapper[4714]: I0129 16:30:21.406860 4714 generic.go:334] "Generic (PLEG): container finished" podID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" containerID="998fc5de34fad62db822ad95298396d20ef6c7ee52c10eeb679ec05818c2f0ea" exitCode=1 Jan 29 16:30:21 crc kubenswrapper[4714]: I0129 16:30:21.406913 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"27df3c21-5ecb-4af0-9e48-a40f826dc75d","Type":"ContainerDied","Data":"998fc5de34fad62db822ad95298396d20ef6c7ee52c10eeb679ec05818c2f0ea"} Jan 29 16:30:21 crc kubenswrapper[4714]: I0129 16:30:21.406963 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"27df3c21-5ecb-4af0-9e48-a40f826dc75d","Type":"ContainerStarted","Data":"28c248e70e5735ed873359efc695cf5ba4233b35ef2d24a816c0f50304261061"} Jan 29 16:30:21 crc kubenswrapper[4714]: I0129 16:30:21.406980 4714 scope.go:117] "RemoveContainer" containerID="4565e725bbc777847d5f682f8f2a88a1faea9eac4df83a536e5f10e10af89617" Jan 29 16:30:21 crc kubenswrapper[4714]: I0129 16:30:21.408428 4714 scope.go:117] "RemoveContainer" containerID="998fc5de34fad62db822ad95298396d20ef6c7ee52c10eeb679ec05818c2f0ea" Jan 29 16:30:21 crc kubenswrapper[4714]: E0129 16:30:21.408769 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(27df3c21-5ecb-4af0-9e48-a40f826dc75d)\"" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" Jan 29 16:30:22 crc kubenswrapper[4714]: I0129 16:30:22.416971 4714 generic.go:334] "Generic (PLEG): container finished" podID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" containerID="28c248e70e5735ed873359efc695cf5ba4233b35ef2d24a816c0f50304261061" exitCode=1 Jan 29 16:30:22 crc kubenswrapper[4714]: I0129 16:30:22.417016 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"27df3c21-5ecb-4af0-9e48-a40f826dc75d","Type":"ContainerDied","Data":"28c248e70e5735ed873359efc695cf5ba4233b35ef2d24a816c0f50304261061"} Jan 29 16:30:22 crc kubenswrapper[4714]: I0129 16:30:22.417054 4714 scope.go:117] "RemoveContainer" containerID="6825612e424e7d18534f6766e6735d779be6e5909c3365b087bf62b9bfd4f305" Jan 29 16:30:22 crc kubenswrapper[4714]: I0129 16:30:22.418542 4714 scope.go:117] "RemoveContainer" containerID="28c248e70e5735ed873359efc695cf5ba4233b35ef2d24a816c0f50304261061" Jan 29 16:30:22 crc kubenswrapper[4714]: I0129 16:30:22.418576 4714 scope.go:117] "RemoveContainer" containerID="998fc5de34fad62db822ad95298396d20ef6c7ee52c10eeb679ec05818c2f0ea" Jan 29 16:30:22 crc kubenswrapper[4714]: E0129 16:30:22.419062 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(27df3c21-5ecb-4af0-9e48-a40f826dc75d)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(27df3c21-5ecb-4af0-9e48-a40f826dc75d)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" Jan 29 16:30:23 crc kubenswrapper[4714]: I0129 16:30:23.430296 4714 scope.go:117] "RemoveContainer" containerID="28c248e70e5735ed873359efc695cf5ba4233b35ef2d24a816c0f50304261061" Jan 29 16:30:23 crc kubenswrapper[4714]: I0129 16:30:23.430353 4714 scope.go:117] "RemoveContainer" containerID="998fc5de34fad62db822ad95298396d20ef6c7ee52c10eeb679ec05818c2f0ea" Jan 29 16:30:23 crc kubenswrapper[4714]: E0129 16:30:23.430834 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(27df3c21-5ecb-4af0-9e48-a40f826dc75d)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(27df3c21-5ecb-4af0-9e48-a40f826dc75d)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" Jan 29 16:30:24 crc kubenswrapper[4714]: E0129 16:30:24.190248 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:30:24 crc kubenswrapper[4714]: I0129 16:30:24.962845 4714 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:24 crc kubenswrapper[4714]: I0129 16:30:24.964113 4714 scope.go:117] "RemoveContainer" containerID="28c248e70e5735ed873359efc695cf5ba4233b35ef2d24a816c0f50304261061" Jan 29 16:30:24 crc kubenswrapper[4714]: I0129 16:30:24.964143 4714 scope.go:117] "RemoveContainer" containerID="998fc5de34fad62db822ad95298396d20ef6c7ee52c10eeb679ec05818c2f0ea" Jan 29 16:30:24 crc kubenswrapper[4714]: E0129 16:30:24.964704 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(27df3c21-5ecb-4af0-9e48-a40f826dc75d)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(27df3c21-5ecb-4af0-9e48-a40f826dc75d)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" Jan 29 16:30:25 crc kubenswrapper[4714]: I0129 16:30:25.962985 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:25 crc kubenswrapper[4714]: I0129 16:30:25.963344 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:25 crc kubenswrapper[4714]: I0129 16:30:25.964102 4714 scope.go:117] "RemoveContainer" containerID="28c248e70e5735ed873359efc695cf5ba4233b35ef2d24a816c0f50304261061" Jan 29 16:30:25 crc kubenswrapper[4714]: I0129 16:30:25.964118 4714 scope.go:117] "RemoveContainer" containerID="998fc5de34fad62db822ad95298396d20ef6c7ee52c10eeb679ec05818c2f0ea" Jan 29 16:30:25 crc kubenswrapper[4714]: E0129 16:30:25.964568 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(27df3c21-5ecb-4af0-9e48-a40f826dc75d)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(27df3c21-5ecb-4af0-9e48-a40f826dc75d)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" Jan 29 16:30:26 crc kubenswrapper[4714]: I0129 16:30:26.133826 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:26 crc kubenswrapper[4714]: I0129 16:30:26.144560 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:27 crc kubenswrapper[4714]: I0129 16:30:27.843912 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:30:27 crc kubenswrapper[4714]: I0129 16:30:27.844291 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:30:28 crc kubenswrapper[4714]: I0129 16:30:28.106788 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.110640 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-zpbtx"] Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.121775 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-zpbtx"] Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.168094 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.168509 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="7c2fde86-e7c8-4605-9750-8464ca4b7d58" containerName="cinder-backup" containerID="cri-o://2caedfb03564966ae2d6b87961e4e7d74fd983fe0c16c41eb96e2d6ca6ebd267" gracePeriod=30 Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.169023 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="7c2fde86-e7c8-4605-9750-8464ca4b7d58" containerName="probe" containerID="cri-o://c236db1e5dbb8d372959347da33330e9e06dab48278781aa11dc0c6a6f372af8" gracePeriod=30 Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.202027 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.214280 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.214551 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="130b46c6-e7e5-4202-bea4-1214ec4766e8" containerName="cinder-scheduler" containerID="cri-o://bccd615573c7599b167665a1108953cb3d1759c6fb3d978bae9a4b7e75fbc11a" gracePeriod=30 Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.214684 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="130b46c6-e7e5-4202-bea4-1214ec4766e8" containerName="probe" containerID="cri-o://bbf7723cee6103ad631b2d89beaf6a72a626280c58f32453fc92b4acddbf7202" gracePeriod=30 Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.228793 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.229057 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="93eb21c7-d0f9-4648-a671-03d3ccd28429" containerName="cinder-api-log" containerID="cri-o://b54fd3f2968e79a20a0b8cb352bdc32d69a2daf77ebdec8342b8fb48a253ab80" gracePeriod=30 Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.229176 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="93eb21c7-d0f9-4648-a671-03d3ccd28429" containerName="cinder-api" containerID="cri-o://0c9e8f216ea8d08fef702164fee1c489d59f45d35804acfb1d8d5c8a08af65d1" gracePeriod=30 Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.234077 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder6e48-account-delete-8kw8k"] Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.235090 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder6e48-account-delete-8kw8k" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.238539 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="cinder-kuttl-tests/cinder-api-0" podUID="93eb21c7-d0f9-4648-a671-03d3ccd28429" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.109:8776/healthcheck\": EOF" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.247467 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder6e48-account-delete-8kw8k"] Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.360963 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2da2589f-6f6a-4921-bc7e-70e1b62979f1-operator-scripts\") pod \"cinder6e48-account-delete-8kw8k\" (UID: \"2da2589f-6f6a-4921-bc7e-70e1b62979f1\") " pod="cinder-kuttl-tests/cinder6e48-account-delete-8kw8k" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.361012 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56bph\" (UniqueName: \"kubernetes.io/projected/2da2589f-6f6a-4921-bc7e-70e1b62979f1-kube-api-access-56bph\") pod \"cinder6e48-account-delete-8kw8k\" (UID: \"2da2589f-6f6a-4921-bc7e-70e1b62979f1\") " pod="cinder-kuttl-tests/cinder6e48-account-delete-8kw8k" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.461888 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2da2589f-6f6a-4921-bc7e-70e1b62979f1-operator-scripts\") pod \"cinder6e48-account-delete-8kw8k\" (UID: \"2da2589f-6f6a-4921-bc7e-70e1b62979f1\") " pod="cinder-kuttl-tests/cinder6e48-account-delete-8kw8k" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.461966 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56bph\" (UniqueName: \"kubernetes.io/projected/2da2589f-6f6a-4921-bc7e-70e1b62979f1-kube-api-access-56bph\") pod \"cinder6e48-account-delete-8kw8k\" (UID: \"2da2589f-6f6a-4921-bc7e-70e1b62979f1\") " pod="cinder-kuttl-tests/cinder6e48-account-delete-8kw8k" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.463055 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2da2589f-6f6a-4921-bc7e-70e1b62979f1-operator-scripts\") pod \"cinder6e48-account-delete-8kw8k\" (UID: \"2da2589f-6f6a-4921-bc7e-70e1b62979f1\") " pod="cinder-kuttl-tests/cinder6e48-account-delete-8kw8k" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.477557 4714 generic.go:334] "Generic (PLEG): container finished" podID="93eb21c7-d0f9-4648-a671-03d3ccd28429" containerID="b54fd3f2968e79a20a0b8cb352bdc32d69a2daf77ebdec8342b8fb48a253ab80" exitCode=143 Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.477610 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"93eb21c7-d0f9-4648-a671-03d3ccd28429","Type":"ContainerDied","Data":"b54fd3f2968e79a20a0b8cb352bdc32d69a2daf77ebdec8342b8fb48a253ab80"} Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.503559 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56bph\" (UniqueName: \"kubernetes.io/projected/2da2589f-6f6a-4921-bc7e-70e1b62979f1-kube-api-access-56bph\") pod \"cinder6e48-account-delete-8kw8k\" (UID: \"2da2589f-6f6a-4921-bc7e-70e1b62979f1\") " pod="cinder-kuttl-tests/cinder6e48-account-delete-8kw8k" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.567686 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.569158 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder6e48-account-delete-8kw8k" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.664947 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-combined-ca-bundle\") pod \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.664999 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-etc-iscsi\") pod \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.665020 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-etc-nvme\") pod \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.665048 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-lib-modules\") pod \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.665122 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-config-data\") pod \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.665154 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-var-locks-brick\") pod \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.665192 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-run\") pod \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.665233 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl96t\" (UniqueName: \"kubernetes.io/projected/27df3c21-5ecb-4af0-9e48-a40f826dc75d-kube-api-access-jl96t\") pod \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.665258 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-etc-machine-id\") pod \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.665277 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-var-lib-cinder\") pod \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.665296 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-dev\") pod \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.665339 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-config-data-custom\") pod \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.665366 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-sys\") pod \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.665405 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-var-locks-cinder\") pod \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.665430 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-scripts\") pod \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\" (UID: \"27df3c21-5ecb-4af0-9e48-a40f826dc75d\") " Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.666619 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "27df3c21-5ecb-4af0-9e48-a40f826dc75d" (UID: "27df3c21-5ecb-4af0-9e48-a40f826dc75d"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.666677 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "27df3c21-5ecb-4af0-9e48-a40f826dc75d" (UID: "27df3c21-5ecb-4af0-9e48-a40f826dc75d"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.666697 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-run" (OuterVolumeSpecName: "run") pod "27df3c21-5ecb-4af0-9e48-a40f826dc75d" (UID: "27df3c21-5ecb-4af0-9e48-a40f826dc75d"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.668174 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "27df3c21-5ecb-4af0-9e48-a40f826dc75d" (UID: "27df3c21-5ecb-4af0-9e48-a40f826dc75d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.668337 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-dev" (OuterVolumeSpecName: "dev") pod "27df3c21-5ecb-4af0-9e48-a40f826dc75d" (UID: "27df3c21-5ecb-4af0-9e48-a40f826dc75d"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.668344 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "27df3c21-5ecb-4af0-9e48-a40f826dc75d" (UID: "27df3c21-5ecb-4af0-9e48-a40f826dc75d"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.668365 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "27df3c21-5ecb-4af0-9e48-a40f826dc75d" (UID: "27df3c21-5ecb-4af0-9e48-a40f826dc75d"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.668377 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-sys" (OuterVolumeSpecName: "sys") pod "27df3c21-5ecb-4af0-9e48-a40f826dc75d" (UID: "27df3c21-5ecb-4af0-9e48-a40f826dc75d"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.668360 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "27df3c21-5ecb-4af0-9e48-a40f826dc75d" (UID: "27df3c21-5ecb-4af0-9e48-a40f826dc75d"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.668398 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "27df3c21-5ecb-4af0-9e48-a40f826dc75d" (UID: "27df3c21-5ecb-4af0-9e48-a40f826dc75d"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.669117 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-scripts" (OuterVolumeSpecName: "scripts") pod "27df3c21-5ecb-4af0-9e48-a40f826dc75d" (UID: "27df3c21-5ecb-4af0-9e48-a40f826dc75d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.671281 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27df3c21-5ecb-4af0-9e48-a40f826dc75d-kube-api-access-jl96t" (OuterVolumeSpecName: "kube-api-access-jl96t") pod "27df3c21-5ecb-4af0-9e48-a40f826dc75d" (UID: "27df3c21-5ecb-4af0-9e48-a40f826dc75d"). InnerVolumeSpecName "kube-api-access-jl96t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.674096 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "27df3c21-5ecb-4af0-9e48-a40f826dc75d" (UID: "27df3c21-5ecb-4af0-9e48-a40f826dc75d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.709342 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27df3c21-5ecb-4af0-9e48-a40f826dc75d" (UID: "27df3c21-5ecb-4af0-9e48-a40f826dc75d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.743613 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-config-data" (OuterVolumeSpecName: "config-data") pod "27df3c21-5ecb-4af0-9e48-a40f826dc75d" (UID: "27df3c21-5ecb-4af0-9e48-a40f826dc75d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.766617 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.766649 4714 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-sys\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.766659 4714 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.766669 4714 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.766676 4714 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.766684 4714 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-etc-nvme\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.766692 4714 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-etc-iscsi\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.766700 4714 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-lib-modules\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.766709 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27df3c21-5ecb-4af0-9e48-a40f826dc75d-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.766719 4714 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-var-locks-brick\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.766726 4714 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-run\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.766733 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl96t\" (UniqueName: \"kubernetes.io/projected/27df3c21-5ecb-4af0-9e48-a40f826dc75d-kube-api-access-jl96t\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.766744 4714 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.766751 4714 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:29 crc kubenswrapper[4714]: I0129 16:30:29.766760 4714 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/27df3c21-5ecb-4af0-9e48-a40f826dc75d-dev\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:30 crc kubenswrapper[4714]: I0129 16:30:30.004471 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder6e48-account-delete-8kw8k"] Jan 29 16:30:30 crc kubenswrapper[4714]: W0129 16:30:30.011064 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2da2589f_6f6a_4921_bc7e_70e1b62979f1.slice/crio-47333c12ea7c2648d34e992cfdd83e276361abff785531f101ec9270b9603aa3 WatchSource:0}: Error finding container 47333c12ea7c2648d34e992cfdd83e276361abff785531f101ec9270b9603aa3: Status 404 returned error can't find the container with id 47333c12ea7c2648d34e992cfdd83e276361abff785531f101ec9270b9603aa3 Jan 29 16:30:30 crc kubenswrapper[4714]: I0129 16:30:30.192285 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cabdde2-0578-405a-9147-efe4d1db7e90" path="/var/lib/kubelet/pods/0cabdde2-0578-405a-9147-efe4d1db7e90/volumes" Jan 29 16:30:30 crc kubenswrapper[4714]: I0129 16:30:30.485718 4714 generic.go:334] "Generic (PLEG): container finished" podID="130b46c6-e7e5-4202-bea4-1214ec4766e8" containerID="bbf7723cee6103ad631b2d89beaf6a72a626280c58f32453fc92b4acddbf7202" exitCode=0 Jan 29 16:30:30 crc kubenswrapper[4714]: I0129 16:30:30.485781 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"130b46c6-e7e5-4202-bea4-1214ec4766e8","Type":"ContainerDied","Data":"bbf7723cee6103ad631b2d89beaf6a72a626280c58f32453fc92b4acddbf7202"} Jan 29 16:30:30 crc kubenswrapper[4714]: I0129 16:30:30.489093 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"27df3c21-5ecb-4af0-9e48-a40f826dc75d","Type":"ContainerDied","Data":"a215b87a9f7bec152d91f39a039625249d594cfc1ffc1ed4fb03f005b6ae19e1"} Jan 29 16:30:30 crc kubenswrapper[4714]: I0129 16:30:30.489150 4714 scope.go:117] "RemoveContainer" containerID="28c248e70e5735ed873359efc695cf5ba4233b35ef2d24a816c0f50304261061" Jan 29 16:30:30 crc kubenswrapper[4714]: I0129 16:30:30.489149 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Jan 29 16:30:30 crc kubenswrapper[4714]: I0129 16:30:30.491335 4714 generic.go:334] "Generic (PLEG): container finished" podID="2da2589f-6f6a-4921-bc7e-70e1b62979f1" containerID="5aeecda1a40201485f4391ac8cf0a5c17c26ea2f9167f27b15c6c783da6f0f44" exitCode=0 Jan 29 16:30:30 crc kubenswrapper[4714]: I0129 16:30:30.491476 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder6e48-account-delete-8kw8k" event={"ID":"2da2589f-6f6a-4921-bc7e-70e1b62979f1","Type":"ContainerDied","Data":"5aeecda1a40201485f4391ac8cf0a5c17c26ea2f9167f27b15c6c783da6f0f44"} Jan 29 16:30:30 crc kubenswrapper[4714]: I0129 16:30:30.491504 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder6e48-account-delete-8kw8k" event={"ID":"2da2589f-6f6a-4921-bc7e-70e1b62979f1","Type":"ContainerStarted","Data":"47333c12ea7c2648d34e992cfdd83e276361abff785531f101ec9270b9603aa3"} Jan 29 16:30:30 crc kubenswrapper[4714]: I0129 16:30:30.496261 4714 generic.go:334] "Generic (PLEG): container finished" podID="7c2fde86-e7c8-4605-9750-8464ca4b7d58" containerID="c236db1e5dbb8d372959347da33330e9e06dab48278781aa11dc0c6a6f372af8" exitCode=0 Jan 29 16:30:30 crc kubenswrapper[4714]: I0129 16:30:30.496350 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"7c2fde86-e7c8-4605-9750-8464ca4b7d58","Type":"ContainerDied","Data":"c236db1e5dbb8d372959347da33330e9e06dab48278781aa11dc0c6a6f372af8"} Jan 29 16:30:30 crc kubenswrapper[4714]: I0129 16:30:30.524168 4714 scope.go:117] "RemoveContainer" containerID="998fc5de34fad62db822ad95298396d20ef6c7ee52c10eeb679ec05818c2f0ea" Jan 29 16:30:30 crc kubenswrapper[4714]: I0129 16:30:30.542269 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Jan 29 16:30:30 crc kubenswrapper[4714]: I0129 16:30:30.548698 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Jan 29 16:30:32 crc kubenswrapper[4714]: I0129 16:30:31.785590 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder6e48-account-delete-8kw8k" Jan 29 16:30:32 crc kubenswrapper[4714]: I0129 16:30:31.899170 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2da2589f-6f6a-4921-bc7e-70e1b62979f1-operator-scripts\") pod \"2da2589f-6f6a-4921-bc7e-70e1b62979f1\" (UID: \"2da2589f-6f6a-4921-bc7e-70e1b62979f1\") " Jan 29 16:30:32 crc kubenswrapper[4714]: I0129 16:30:31.899266 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56bph\" (UniqueName: \"kubernetes.io/projected/2da2589f-6f6a-4921-bc7e-70e1b62979f1-kube-api-access-56bph\") pod \"2da2589f-6f6a-4921-bc7e-70e1b62979f1\" (UID: \"2da2589f-6f6a-4921-bc7e-70e1b62979f1\") " Jan 29 16:30:32 crc kubenswrapper[4714]: I0129 16:30:31.900054 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2da2589f-6f6a-4921-bc7e-70e1b62979f1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2da2589f-6f6a-4921-bc7e-70e1b62979f1" (UID: "2da2589f-6f6a-4921-bc7e-70e1b62979f1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:30:32 crc kubenswrapper[4714]: I0129 16:30:31.912259 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2da2589f-6f6a-4921-bc7e-70e1b62979f1-kube-api-access-56bph" (OuterVolumeSpecName: "kube-api-access-56bph") pod "2da2589f-6f6a-4921-bc7e-70e1b62979f1" (UID: "2da2589f-6f6a-4921-bc7e-70e1b62979f1"). InnerVolumeSpecName "kube-api-access-56bph". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:32 crc kubenswrapper[4714]: I0129 16:30:32.001447 4714 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2da2589f-6f6a-4921-bc7e-70e1b62979f1-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:32 crc kubenswrapper[4714]: I0129 16:30:32.001529 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56bph\" (UniqueName: \"kubernetes.io/projected/2da2589f-6f6a-4921-bc7e-70e1b62979f1-kube-api-access-56bph\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:32 crc kubenswrapper[4714]: I0129 16:30:32.196261 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" path="/var/lib/kubelet/pods/27df3c21-5ecb-4af0-9e48-a40f826dc75d/volumes" Jan 29 16:30:32 crc kubenswrapper[4714]: I0129 16:30:32.520012 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder6e48-account-delete-8kw8k" event={"ID":"2da2589f-6f6a-4921-bc7e-70e1b62979f1","Type":"ContainerDied","Data":"47333c12ea7c2648d34e992cfdd83e276361abff785531f101ec9270b9603aa3"} Jan 29 16:30:32 crc kubenswrapper[4714]: I0129 16:30:32.520073 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47333c12ea7c2648d34e992cfdd83e276361abff785531f101ec9270b9603aa3" Jan 29 16:30:32 crc kubenswrapper[4714]: I0129 16:30:32.520155 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder6e48-account-delete-8kw8k" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.527679 4714 generic.go:334] "Generic (PLEG): container finished" podID="7c2fde86-e7c8-4605-9750-8464ca4b7d58" containerID="2caedfb03564966ae2d6b87961e4e7d74fd983fe0c16c41eb96e2d6ca6ebd267" exitCode=0 Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.527786 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"7c2fde86-e7c8-4605-9750-8464ca4b7d58","Type":"ContainerDied","Data":"2caedfb03564966ae2d6b87961e4e7d74fd983fe0c16c41eb96e2d6ca6ebd267"} Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.529719 4714 generic.go:334] "Generic (PLEG): container finished" podID="130b46c6-e7e5-4202-bea4-1214ec4766e8" containerID="bccd615573c7599b167665a1108953cb3d1759c6fb3d978bae9a4b7e75fbc11a" exitCode=0 Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.529758 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"130b46c6-e7e5-4202-bea4-1214ec4766e8","Type":"ContainerDied","Data":"bccd615573c7599b167665a1108953cb3d1759c6fb3d978bae9a4b7e75fbc11a"} Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.651798 4714 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/cinder-api-0" podUID="93eb21c7-d0f9-4648-a671-03d3ccd28429" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.109:8776/healthcheck\": read tcp 10.217.0.2:56774->10.217.0.109:8776: read: connection reset by peer" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.659571 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.727647 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-config-data\") pod \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.727681 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-combined-ca-bundle\") pod \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.727716 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-lib-modules\") pod \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.727740 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-dev\") pod \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.727761 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-sys\") pod \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.727801 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-etc-iscsi\") pod \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.727882 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "7c2fde86-e7c8-4605-9750-8464ca4b7d58" (UID: "7c2fde86-e7c8-4605-9750-8464ca4b7d58"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.731547 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-var-locks-brick\") pod \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.731880 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-run\") pod \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.731913 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bl9k4\" (UniqueName: \"kubernetes.io/projected/7c2fde86-e7c8-4605-9750-8464ca4b7d58-kube-api-access-bl9k4\") pod \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.731634 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-dev" (OuterVolumeSpecName: "dev") pod "7c2fde86-e7c8-4605-9750-8464ca4b7d58" (UID: "7c2fde86-e7c8-4605-9750-8464ca4b7d58"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.731669 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-sys" (OuterVolumeSpecName: "sys") pod "7c2fde86-e7c8-4605-9750-8464ca4b7d58" (UID: "7c2fde86-e7c8-4605-9750-8464ca4b7d58"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.731690 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "7c2fde86-e7c8-4605-9750-8464ca4b7d58" (UID: "7c2fde86-e7c8-4605-9750-8464ca4b7d58"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.731990 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-run" (OuterVolumeSpecName: "run") pod "7c2fde86-e7c8-4605-9750-8464ca4b7d58" (UID: "7c2fde86-e7c8-4605-9750-8464ca4b7d58"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.731741 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "7c2fde86-e7c8-4605-9750-8464ca4b7d58" (UID: "7c2fde86-e7c8-4605-9750-8464ca4b7d58"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.731951 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-scripts\") pod \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.732089 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-etc-machine-id\") pod \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.732132 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-config-data-custom\") pod \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.732157 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-var-locks-cinder\") pod \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.732202 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-etc-nvme\") pod \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.732230 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-var-lib-cinder\") pod \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\" (UID: \"7c2fde86-e7c8-4605-9750-8464ca4b7d58\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.732690 4714 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-lib-modules\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.732707 4714 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-dev\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.732717 4714 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-sys\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.732726 4714 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-etc-iscsi\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.732734 4714 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-var-locks-brick\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.732743 4714 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-run\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.732771 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "7c2fde86-e7c8-4605-9750-8464ca4b7d58" (UID: "7c2fde86-e7c8-4605-9750-8464ca4b7d58"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.732791 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7c2fde86-e7c8-4605-9750-8464ca4b7d58" (UID: "7c2fde86-e7c8-4605-9750-8464ca4b7d58"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.733173 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "7c2fde86-e7c8-4605-9750-8464ca4b7d58" (UID: "7c2fde86-e7c8-4605-9750-8464ca4b7d58"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.733204 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "7c2fde86-e7c8-4605-9750-8464ca4b7d58" (UID: "7c2fde86-e7c8-4605-9750-8464ca4b7d58"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.735508 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c2fde86-e7c8-4605-9750-8464ca4b7d58-kube-api-access-bl9k4" (OuterVolumeSpecName: "kube-api-access-bl9k4") pod "7c2fde86-e7c8-4605-9750-8464ca4b7d58" (UID: "7c2fde86-e7c8-4605-9750-8464ca4b7d58"). InnerVolumeSpecName "kube-api-access-bl9k4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.735695 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-scripts" (OuterVolumeSpecName: "scripts") pod "7c2fde86-e7c8-4605-9750-8464ca4b7d58" (UID: "7c2fde86-e7c8-4605-9750-8464ca4b7d58"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.736087 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7c2fde86-e7c8-4605-9750-8464ca4b7d58" (UID: "7c2fde86-e7c8-4605-9750-8464ca4b7d58"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.753820 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.774508 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7c2fde86-e7c8-4605-9750-8464ca4b7d58" (UID: "7c2fde86-e7c8-4605-9750-8464ca4b7d58"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.799140 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-config-data" (OuterVolumeSpecName: "config-data") pod "7c2fde86-e7c8-4605-9750-8464ca4b7d58" (UID: "7c2fde86-e7c8-4605-9750-8464ca4b7d58"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.834072 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-config-data-custom\") pod \"130b46c6-e7e5-4202-bea4-1214ec4766e8\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.834175 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-scripts\") pod \"130b46c6-e7e5-4202-bea4-1214ec4766e8\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.834266 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-combined-ca-bundle\") pod \"130b46c6-e7e5-4202-bea4-1214ec4766e8\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.834306 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-config-data\") pod \"130b46c6-e7e5-4202-bea4-1214ec4766e8\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.834334 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85zz9\" (UniqueName: \"kubernetes.io/projected/130b46c6-e7e5-4202-bea4-1214ec4766e8-kube-api-access-85zz9\") pod \"130b46c6-e7e5-4202-bea4-1214ec4766e8\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.834358 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/130b46c6-e7e5-4202-bea4-1214ec4766e8-etc-machine-id\") pod \"130b46c6-e7e5-4202-bea4-1214ec4766e8\" (UID: \"130b46c6-e7e5-4202-bea4-1214ec4766e8\") " Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.834777 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.834796 4714 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.834806 4714 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-etc-nvme\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.834815 4714 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.834824 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.834833 4714 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.834844 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bl9k4\" (UniqueName: \"kubernetes.io/projected/7c2fde86-e7c8-4605-9750-8464ca4b7d58-kube-api-access-bl9k4\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.834854 4714 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c2fde86-e7c8-4605-9750-8464ca4b7d58-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.834863 4714 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7c2fde86-e7c8-4605-9750-8464ca4b7d58-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.835269 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/130b46c6-e7e5-4202-bea4-1214ec4766e8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "130b46c6-e7e5-4202-bea4-1214ec4766e8" (UID: "130b46c6-e7e5-4202-bea4-1214ec4766e8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.838274 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "130b46c6-e7e5-4202-bea4-1214ec4766e8" (UID: "130b46c6-e7e5-4202-bea4-1214ec4766e8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.838373 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-scripts" (OuterVolumeSpecName: "scripts") pod "130b46c6-e7e5-4202-bea4-1214ec4766e8" (UID: "130b46c6-e7e5-4202-bea4-1214ec4766e8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.840535 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/130b46c6-e7e5-4202-bea4-1214ec4766e8-kube-api-access-85zz9" (OuterVolumeSpecName: "kube-api-access-85zz9") pod "130b46c6-e7e5-4202-bea4-1214ec4766e8" (UID: "130b46c6-e7e5-4202-bea4-1214ec4766e8"). InnerVolumeSpecName "kube-api-access-85zz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.880625 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "130b46c6-e7e5-4202-bea4-1214ec4766e8" (UID: "130b46c6-e7e5-4202-bea4-1214ec4766e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.901089 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-config-data" (OuterVolumeSpecName: "config-data") pod "130b46c6-e7e5-4202-bea4-1214ec4766e8" (UID: "130b46c6-e7e5-4202-bea4-1214ec4766e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.935835 4714 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.935868 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.935877 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85zz9\" (UniqueName: \"kubernetes.io/projected/130b46c6-e7e5-4202-bea4-1214ec4766e8-kube-api-access-85zz9\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.935889 4714 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/130b46c6-e7e5-4202-bea4-1214ec4766e8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.935898 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.935906 4714 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/130b46c6-e7e5-4202-bea4-1214ec4766e8-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:33 crc kubenswrapper[4714]: I0129 16:30:33.994314 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.037207 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-combined-ca-bundle\") pod \"93eb21c7-d0f9-4648-a671-03d3ccd28429\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.037277 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-scripts\") pod \"93eb21c7-d0f9-4648-a671-03d3ccd28429\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.037332 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93eb21c7-d0f9-4648-a671-03d3ccd28429-logs\") pod \"93eb21c7-d0f9-4648-a671-03d3ccd28429\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.037379 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/93eb21c7-d0f9-4648-a671-03d3ccd28429-etc-machine-id\") pod \"93eb21c7-d0f9-4648-a671-03d3ccd28429\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.037402 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-internal-tls-certs\") pod \"93eb21c7-d0f9-4648-a671-03d3ccd28429\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.037482 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-str9z\" (UniqueName: \"kubernetes.io/projected/93eb21c7-d0f9-4648-a671-03d3ccd28429-kube-api-access-str9z\") pod \"93eb21c7-d0f9-4648-a671-03d3ccd28429\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.037473 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93eb21c7-d0f9-4648-a671-03d3ccd28429-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "93eb21c7-d0f9-4648-a671-03d3ccd28429" (UID: "93eb21c7-d0f9-4648-a671-03d3ccd28429"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.038277 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93eb21c7-d0f9-4648-a671-03d3ccd28429-logs" (OuterVolumeSpecName: "logs") pod "93eb21c7-d0f9-4648-a671-03d3ccd28429" (UID: "93eb21c7-d0f9-4648-a671-03d3ccd28429"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.038296 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-public-tls-certs\") pod \"93eb21c7-d0f9-4648-a671-03d3ccd28429\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.038437 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-config-data\") pod \"93eb21c7-d0f9-4648-a671-03d3ccd28429\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.038518 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-config-data-custom\") pod \"93eb21c7-d0f9-4648-a671-03d3ccd28429\" (UID: \"93eb21c7-d0f9-4648-a671-03d3ccd28429\") " Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.039119 4714 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93eb21c7-d0f9-4648-a671-03d3ccd28429-logs\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.039147 4714 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/93eb21c7-d0f9-4648-a671-03d3ccd28429-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.040617 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93eb21c7-d0f9-4648-a671-03d3ccd28429-kube-api-access-str9z" (OuterVolumeSpecName: "kube-api-access-str9z") pod "93eb21c7-d0f9-4648-a671-03d3ccd28429" (UID: "93eb21c7-d0f9-4648-a671-03d3ccd28429"). InnerVolumeSpecName "kube-api-access-str9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.041405 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-scripts" (OuterVolumeSpecName: "scripts") pod "93eb21c7-d0f9-4648-a671-03d3ccd28429" (UID: "93eb21c7-d0f9-4648-a671-03d3ccd28429"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.041510 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "93eb21c7-d0f9-4648-a671-03d3ccd28429" (UID: "93eb21c7-d0f9-4648-a671-03d3ccd28429"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.058174 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93eb21c7-d0f9-4648-a671-03d3ccd28429" (UID: "93eb21c7-d0f9-4648-a671-03d3ccd28429"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.073106 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-config-data" (OuterVolumeSpecName: "config-data") pod "93eb21c7-d0f9-4648-a671-03d3ccd28429" (UID: "93eb21c7-d0f9-4648-a671-03d3ccd28429"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.075252 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "93eb21c7-d0f9-4648-a671-03d3ccd28429" (UID: "93eb21c7-d0f9-4648-a671-03d3ccd28429"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.079393 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "93eb21c7-d0f9-4648-a671-03d3ccd28429" (UID: "93eb21c7-d0f9-4648-a671-03d3ccd28429"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.140674 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-str9z\" (UniqueName: \"kubernetes.io/projected/93eb21c7-d0f9-4648-a671-03d3ccd28429-kube-api-access-str9z\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.140712 4714 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.140722 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.140732 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.140740 4714 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.140748 4714 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.140757 4714 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/93eb21c7-d0f9-4648-a671-03d3ccd28429-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.244145 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-hqbqv"] Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.251890 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-hqbqv"] Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.270392 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder6e48-account-delete-8kw8k"] Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.283314 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-6e48-account-create-update-qff9j"] Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.295713 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-6e48-account-create-update-qff9j"] Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.307985 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder6e48-account-delete-8kw8k"] Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.541869 4714 generic.go:334] "Generic (PLEG): container finished" podID="93eb21c7-d0f9-4648-a671-03d3ccd28429" containerID="0c9e8f216ea8d08fef702164fee1c489d59f45d35804acfb1d8d5c8a08af65d1" exitCode=0 Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.541919 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"93eb21c7-d0f9-4648-a671-03d3ccd28429","Type":"ContainerDied","Data":"0c9e8f216ea8d08fef702164fee1c489d59f45d35804acfb1d8d5c8a08af65d1"} Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.541973 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.542011 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"93eb21c7-d0f9-4648-a671-03d3ccd28429","Type":"ContainerDied","Data":"9cf9a3f95739dab5cf604ede39f2741a3a785d75007c516172ae75ff6b3665b1"} Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.542050 4714 scope.go:117] "RemoveContainer" containerID="0c9e8f216ea8d08fef702164fee1c489d59f45d35804acfb1d8d5c8a08af65d1" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.545549 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"7c2fde86-e7c8-4605-9750-8464ca4b7d58","Type":"ContainerDied","Data":"e25204d76e2148fee59f9d4ff12c3c8594f24cfcc42364e257052da501eeec8a"} Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.545563 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.549386 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"130b46c6-e7e5-4202-bea4-1214ec4766e8","Type":"ContainerDied","Data":"4117860f9b4009b5187f645af44ec4d5a39f7166f668190572db33d3a8f8e24c"} Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.549505 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.566440 4714 scope.go:117] "RemoveContainer" containerID="b54fd3f2968e79a20a0b8cb352bdc32d69a2daf77ebdec8342b8fb48a253ab80" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.577590 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.584468 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.587984 4714 scope.go:117] "RemoveContainer" containerID="0c9e8f216ea8d08fef702164fee1c489d59f45d35804acfb1d8d5c8a08af65d1" Jan 29 16:30:34 crc kubenswrapper[4714]: E0129 16:30:34.588492 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c9e8f216ea8d08fef702164fee1c489d59f45d35804acfb1d8d5c8a08af65d1\": container with ID starting with 0c9e8f216ea8d08fef702164fee1c489d59f45d35804acfb1d8d5c8a08af65d1 not found: ID does not exist" containerID="0c9e8f216ea8d08fef702164fee1c489d59f45d35804acfb1d8d5c8a08af65d1" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.588541 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c9e8f216ea8d08fef702164fee1c489d59f45d35804acfb1d8d5c8a08af65d1"} err="failed to get container status \"0c9e8f216ea8d08fef702164fee1c489d59f45d35804acfb1d8d5c8a08af65d1\": rpc error: code = NotFound desc = could not find container \"0c9e8f216ea8d08fef702164fee1c489d59f45d35804acfb1d8d5c8a08af65d1\": container with ID starting with 0c9e8f216ea8d08fef702164fee1c489d59f45d35804acfb1d8d5c8a08af65d1 not found: ID does not exist" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.588580 4714 scope.go:117] "RemoveContainer" containerID="b54fd3f2968e79a20a0b8cb352bdc32d69a2daf77ebdec8342b8fb48a253ab80" Jan 29 16:30:34 crc kubenswrapper[4714]: E0129 16:30:34.588873 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b54fd3f2968e79a20a0b8cb352bdc32d69a2daf77ebdec8342b8fb48a253ab80\": container with ID starting with b54fd3f2968e79a20a0b8cb352bdc32d69a2daf77ebdec8342b8fb48a253ab80 not found: ID does not exist" containerID="b54fd3f2968e79a20a0b8cb352bdc32d69a2daf77ebdec8342b8fb48a253ab80" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.588902 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b54fd3f2968e79a20a0b8cb352bdc32d69a2daf77ebdec8342b8fb48a253ab80"} err="failed to get container status \"b54fd3f2968e79a20a0b8cb352bdc32d69a2daf77ebdec8342b8fb48a253ab80\": rpc error: code = NotFound desc = could not find container \"b54fd3f2968e79a20a0b8cb352bdc32d69a2daf77ebdec8342b8fb48a253ab80\": container with ID starting with b54fd3f2968e79a20a0b8cb352bdc32d69a2daf77ebdec8342b8fb48a253ab80 not found: ID does not exist" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.588920 4714 scope.go:117] "RemoveContainer" containerID="c236db1e5dbb8d372959347da33330e9e06dab48278781aa11dc0c6a6f372af8" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.597053 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.601515 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.606351 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.608898 4714 scope.go:117] "RemoveContainer" containerID="2caedfb03564966ae2d6b87961e4e7d74fd983fe0c16c41eb96e2d6ca6ebd267" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.610770 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.624001 4714 scope.go:117] "RemoveContainer" containerID="bbf7723cee6103ad631b2d89beaf6a72a626280c58f32453fc92b4acddbf7202" Jan 29 16:30:34 crc kubenswrapper[4714]: I0129 16:30:34.637769 4714 scope.go:117] "RemoveContainer" containerID="bccd615573c7599b167665a1108953cb3d1759c6fb3d978bae9a4b7e75fbc11a" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.673844 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-6xxc6"] Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.683898 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-6xxc6"] Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.690190 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-kj27d"] Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.697713 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-kj27d"] Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.704029 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-db9b49999-6gd95"] Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.704276 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" podUID="fc06a535-6f60-438e-b52d-5dc90fae8c67" containerName="keystone-api" containerID="cri-o://4adeb441736bb7d27c549844b9147979d27f5dbed7e42a9b990619d275a00000" gracePeriod=30 Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739322 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone52f2-account-delete-qxsl4"] Jan 29 16:30:35 crc kubenswrapper[4714]: E0129 16:30:35.739557 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2da2589f-6f6a-4921-bc7e-70e1b62979f1" containerName="mariadb-account-delete" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739571 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="2da2589f-6f6a-4921-bc7e-70e1b62979f1" containerName="mariadb-account-delete" Jan 29 16:30:35 crc kubenswrapper[4714]: E0129 16:30:35.739583 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" containerName="probe" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739589 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" containerName="probe" Jan 29 16:30:35 crc kubenswrapper[4714]: E0129 16:30:35.739597 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93eb21c7-d0f9-4648-a671-03d3ccd28429" containerName="cinder-api" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739602 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="93eb21c7-d0f9-4648-a671-03d3ccd28429" containerName="cinder-api" Jan 29 16:30:35 crc kubenswrapper[4714]: E0129 16:30:35.739612 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93eb21c7-d0f9-4648-a671-03d3ccd28429" containerName="cinder-api-log" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739618 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="93eb21c7-d0f9-4648-a671-03d3ccd28429" containerName="cinder-api-log" Jan 29 16:30:35 crc kubenswrapper[4714]: E0129 16:30:35.739629 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" containerName="cinder-volume" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739635 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" containerName="cinder-volume" Jan 29 16:30:35 crc kubenswrapper[4714]: E0129 16:30:35.739647 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" containerName="probe" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739653 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" containerName="probe" Jan 29 16:30:35 crc kubenswrapper[4714]: E0129 16:30:35.739661 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c2fde86-e7c8-4605-9750-8464ca4b7d58" containerName="cinder-backup" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739666 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c2fde86-e7c8-4605-9750-8464ca4b7d58" containerName="cinder-backup" Jan 29 16:30:35 crc kubenswrapper[4714]: E0129 16:30:35.739672 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="130b46c6-e7e5-4202-bea4-1214ec4766e8" containerName="probe" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739678 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="130b46c6-e7e5-4202-bea4-1214ec4766e8" containerName="probe" Jan 29 16:30:35 crc kubenswrapper[4714]: E0129 16:30:35.739686 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="130b46c6-e7e5-4202-bea4-1214ec4766e8" containerName="cinder-scheduler" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739693 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="130b46c6-e7e5-4202-bea4-1214ec4766e8" containerName="cinder-scheduler" Jan 29 16:30:35 crc kubenswrapper[4714]: E0129 16:30:35.739701 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c2fde86-e7c8-4605-9750-8464ca4b7d58" containerName="probe" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739707 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c2fde86-e7c8-4605-9750-8464ca4b7d58" containerName="probe" Jan 29 16:30:35 crc kubenswrapper[4714]: E0129 16:30:35.739714 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" containerName="cinder-volume" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739722 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" containerName="cinder-volume" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739817 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" containerName="probe" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739826 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="93eb21c7-d0f9-4648-a671-03d3ccd28429" containerName="cinder-api-log" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739835 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="93eb21c7-d0f9-4648-a671-03d3ccd28429" containerName="cinder-api" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739841 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c2fde86-e7c8-4605-9750-8464ca4b7d58" containerName="cinder-backup" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739852 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" containerName="cinder-volume" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739859 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c2fde86-e7c8-4605-9750-8464ca4b7d58" containerName="probe" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739866 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="130b46c6-e7e5-4202-bea4-1214ec4766e8" containerName="probe" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739875 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="2da2589f-6f6a-4921-bc7e-70e1b62979f1" containerName="mariadb-account-delete" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739882 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" containerName="cinder-volume" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.739890 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="130b46c6-e7e5-4202-bea4-1214ec4766e8" containerName="cinder-scheduler" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.740500 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone52f2-account-delete-qxsl4" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.751115 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone52f2-account-delete-qxsl4"] Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.869550 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/112a1cde-5990-4140-97dd-c2bbc4f73197-operator-scripts\") pod \"keystone52f2-account-delete-qxsl4\" (UID: \"112a1cde-5990-4140-97dd-c2bbc4f73197\") " pod="cinder-kuttl-tests/keystone52f2-account-delete-qxsl4" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.869615 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrfh5\" (UniqueName: \"kubernetes.io/projected/112a1cde-5990-4140-97dd-c2bbc4f73197-kube-api-access-rrfh5\") pod \"keystone52f2-account-delete-qxsl4\" (UID: \"112a1cde-5990-4140-97dd-c2bbc4f73197\") " pod="cinder-kuttl-tests/keystone52f2-account-delete-qxsl4" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.971410 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/112a1cde-5990-4140-97dd-c2bbc4f73197-operator-scripts\") pod \"keystone52f2-account-delete-qxsl4\" (UID: \"112a1cde-5990-4140-97dd-c2bbc4f73197\") " pod="cinder-kuttl-tests/keystone52f2-account-delete-qxsl4" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.971468 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrfh5\" (UniqueName: \"kubernetes.io/projected/112a1cde-5990-4140-97dd-c2bbc4f73197-kube-api-access-rrfh5\") pod \"keystone52f2-account-delete-qxsl4\" (UID: \"112a1cde-5990-4140-97dd-c2bbc4f73197\") " pod="cinder-kuttl-tests/keystone52f2-account-delete-qxsl4" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.972254 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/112a1cde-5990-4140-97dd-c2bbc4f73197-operator-scripts\") pod \"keystone52f2-account-delete-qxsl4\" (UID: \"112a1cde-5990-4140-97dd-c2bbc4f73197\") " pod="cinder-kuttl-tests/keystone52f2-account-delete-qxsl4" Jan 29 16:30:35 crc kubenswrapper[4714]: I0129 16:30:35.995315 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrfh5\" (UniqueName: \"kubernetes.io/projected/112a1cde-5990-4140-97dd-c2bbc4f73197-kube-api-access-rrfh5\") pod \"keystone52f2-account-delete-qxsl4\" (UID: \"112a1cde-5990-4140-97dd-c2bbc4f73197\") " pod="cinder-kuttl-tests/keystone52f2-account-delete-qxsl4" Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.065893 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone52f2-account-delete-qxsl4" Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.196576 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="130b46c6-e7e5-4202-bea4-1214ec4766e8" path="/var/lib/kubelet/pods/130b46c6-e7e5-4202-bea4-1214ec4766e8/volumes" Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.198191 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2da2589f-6f6a-4921-bc7e-70e1b62979f1" path="/var/lib/kubelet/pods/2da2589f-6f6a-4921-bc7e-70e1b62979f1/volumes" Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.199329 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e9c54b1-972a-4807-90af-f94a884002bd" path="/var/lib/kubelet/pods/5e9c54b1-972a-4807-90af-f94a884002bd/volumes" Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.201149 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e33ba3d-9561-441b-b835-fbdb6ce97d23" path="/var/lib/kubelet/pods/6e33ba3d-9561-441b-b835-fbdb6ce97d23/volumes" Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.202312 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c2fde86-e7c8-4605-9750-8464ca4b7d58" path="/var/lib/kubelet/pods/7c2fde86-e7c8-4605-9750-8464ca4b7d58/volumes" Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.203510 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93eb21c7-d0f9-4648-a671-03d3ccd28429" path="/var/lib/kubelet/pods/93eb21c7-d0f9-4648-a671-03d3ccd28429/volumes" Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.205104 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3306850-8843-48e1-b203-7f52de72682f" path="/var/lib/kubelet/pods/c3306850-8843-48e1-b203-7f52de72682f/volumes" Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.205778 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3e03982-d953-488f-a01a-5024f64ad7da" path="/var/lib/kubelet/pods/f3e03982-d953-488f-a01a-5024f64ad7da/volumes" Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.501871 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone52f2-account-delete-qxsl4"] Jan 29 16:30:36 crc kubenswrapper[4714]: W0129 16:30:36.511038 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod112a1cde_5990_4140_97dd_c2bbc4f73197.slice/crio-886af71ec9bb3cd0c3ff50ec50762de0bc190d913b4893fee8dbe7877dbfc19b WatchSource:0}: Error finding container 886af71ec9bb3cd0c3ff50ec50762de0bc190d913b4893fee8dbe7877dbfc19b: Status 404 returned error can't find the container with id 886af71ec9bb3cd0c3ff50ec50762de0bc190d913b4893fee8dbe7877dbfc19b Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.581033 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/root-account-create-update-2fh2r"] Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.618503 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/root-account-create-update-2fh2r"] Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.620361 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone52f2-account-delete-qxsl4" event={"ID":"112a1cde-5990-4140-97dd-c2bbc4f73197","Type":"ContainerStarted","Data":"886af71ec9bb3cd0c3ff50ec50762de0bc190d913b4893fee8dbe7877dbfc19b"} Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.632103 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/root-account-create-update-jg6sh"] Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.632527 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="27df3c21-5ecb-4af0-9e48-a40f826dc75d" containerName="probe" Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.633049 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/root-account-create-update-jg6sh" Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.636061 4714 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"openstack-mariadb-root-db-secret" Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.638979 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.660561 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.665919 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.671082 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/root-account-create-update-jg6sh"] Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.684783 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc4cd\" (UniqueName: \"kubernetes.io/projected/4a2536dd-4258-4b5b-863c-a76431c992ee-kube-api-access-rc4cd\") pod \"root-account-create-update-jg6sh\" (UID: \"4a2536dd-4258-4b5b-863c-a76431c992ee\") " pod="cinder-kuttl-tests/root-account-create-update-jg6sh" Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.684832 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a2536dd-4258-4b5b-863c-a76431c992ee-operator-scripts\") pod \"root-account-create-update-jg6sh\" (UID: \"4a2536dd-4258-4b5b-863c-a76431c992ee\") " pod="cinder-kuttl-tests/root-account-create-update-jg6sh" Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.687463 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/root-account-create-update-jg6sh"] Jan 29 16:30:36 crc kubenswrapper[4714]: E0129 16:30:36.687834 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-rc4cd operator-scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="cinder-kuttl-tests/root-account-create-update-jg6sh" podUID="4a2536dd-4258-4b5b-863c-a76431c992ee" Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.786616 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc4cd\" (UniqueName: \"kubernetes.io/projected/4a2536dd-4258-4b5b-863c-a76431c992ee-kube-api-access-rc4cd\") pod \"root-account-create-update-jg6sh\" (UID: \"4a2536dd-4258-4b5b-863c-a76431c992ee\") " pod="cinder-kuttl-tests/root-account-create-update-jg6sh" Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.786670 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a2536dd-4258-4b5b-863c-a76431c992ee-operator-scripts\") pod \"root-account-create-update-jg6sh\" (UID: \"4a2536dd-4258-4b5b-863c-a76431c992ee\") " pod="cinder-kuttl-tests/root-account-create-update-jg6sh" Jan 29 16:30:36 crc kubenswrapper[4714]: E0129 16:30:36.786836 4714 configmap.go:193] Couldn't get configMap cinder-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Jan 29 16:30:36 crc kubenswrapper[4714]: E0129 16:30:36.786914 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4a2536dd-4258-4b5b-863c-a76431c992ee-operator-scripts podName:4a2536dd-4258-4b5b-863c-a76431c992ee nodeName:}" failed. No retries permitted until 2026-01-29 16:30:37.286895536 +0000 UTC m=+1243.807396656 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4a2536dd-4258-4b5b-863c-a76431c992ee-operator-scripts") pod "root-account-create-update-jg6sh" (UID: "4a2536dd-4258-4b5b-863c-a76431c992ee") : configmap "openstack-scripts" not found Jan 29 16:30:36 crc kubenswrapper[4714]: I0129 16:30:36.787477 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/openstack-galera-2" podUID="e367e739-45d9-4c71-82fa-ecda02da3277" containerName="galera" containerID="cri-o://6653bc0073fbdde40468fb06b953c1302846763f2888418b4fc21509ef3915d2" gracePeriod=30 Jan 29 16:30:36 crc kubenswrapper[4714]: E0129 16:30:36.790611 4714 projected.go:194] Error preparing data for projected volume kube-api-access-rc4cd for pod cinder-kuttl-tests/root-account-create-update-jg6sh: failed to fetch token: serviceaccounts "galera-openstack" not found Jan 29 16:30:36 crc kubenswrapper[4714]: E0129 16:30:36.790655 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4a2536dd-4258-4b5b-863c-a76431c992ee-kube-api-access-rc4cd podName:4a2536dd-4258-4b5b-863c-a76431c992ee nodeName:}" failed. No retries permitted until 2026-01-29 16:30:37.290644034 +0000 UTC m=+1243.811145154 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-rc4cd" (UniqueName: "kubernetes.io/projected/4a2536dd-4258-4b5b-863c-a76431c992ee-kube-api-access-rc4cd") pod "root-account-create-update-jg6sh" (UID: "4a2536dd-4258-4b5b-863c-a76431c992ee") : failed to fetch token: serviceaccounts "galera-openstack" not found Jan 29 16:30:37 crc kubenswrapper[4714]: E0129 16:30:37.185750 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.190019 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.190335 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/memcached-0" podUID="d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea" containerName="memcached" containerID="cri-o://5244ecc09d3c74f3a195f77af7759f9d242093032f09d5208383b4619dda295d" gracePeriod=30 Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.294200 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc4cd\" (UniqueName: \"kubernetes.io/projected/4a2536dd-4258-4b5b-863c-a76431c992ee-kube-api-access-rc4cd\") pod \"root-account-create-update-jg6sh\" (UID: \"4a2536dd-4258-4b5b-863c-a76431c992ee\") " pod="cinder-kuttl-tests/root-account-create-update-jg6sh" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.294277 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a2536dd-4258-4b5b-863c-a76431c992ee-operator-scripts\") pod \"root-account-create-update-jg6sh\" (UID: \"4a2536dd-4258-4b5b-863c-a76431c992ee\") " pod="cinder-kuttl-tests/root-account-create-update-jg6sh" Jan 29 16:30:37 crc kubenswrapper[4714]: E0129 16:30:37.295309 4714 configmap.go:193] Couldn't get configMap cinder-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Jan 29 16:30:37 crc kubenswrapper[4714]: E0129 16:30:37.295365 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4a2536dd-4258-4b5b-863c-a76431c992ee-operator-scripts podName:4a2536dd-4258-4b5b-863c-a76431c992ee nodeName:}" failed. No retries permitted until 2026-01-29 16:30:38.295348663 +0000 UTC m=+1244.815849783 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4a2536dd-4258-4b5b-863c-a76431c992ee-operator-scripts") pod "root-account-create-update-jg6sh" (UID: "4a2536dd-4258-4b5b-863c-a76431c992ee") : configmap "openstack-scripts" not found Jan 29 16:30:37 crc kubenswrapper[4714]: E0129 16:30:37.299066 4714 projected.go:194] Error preparing data for projected volume kube-api-access-rc4cd for pod cinder-kuttl-tests/root-account-create-update-jg6sh: failed to fetch token: serviceaccounts "galera-openstack" not found Jan 29 16:30:37 crc kubenswrapper[4714]: E0129 16:30:37.299170 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4a2536dd-4258-4b5b-863c-a76431c992ee-kube-api-access-rc4cd podName:4a2536dd-4258-4b5b-863c-a76431c992ee nodeName:}" failed. No retries permitted until 2026-01-29 16:30:38.299144241 +0000 UTC m=+1244.819645381 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-rc4cd" (UniqueName: "kubernetes.io/projected/4a2536dd-4258-4b5b-863c-a76431c992ee-kube-api-access-rc4cd") pod "root-account-create-update-jg6sh" (UID: "4a2536dd-4258-4b5b-863c-a76431c992ee") : failed to fetch token: serviceaccounts "galera-openstack" not found Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.601518 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.607077 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.641159 4714 generic.go:334] "Generic (PLEG): container finished" podID="e367e739-45d9-4c71-82fa-ecda02da3277" containerID="6653bc0073fbdde40468fb06b953c1302846763f2888418b4fc21509ef3915d2" exitCode=0 Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.641220 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"e367e739-45d9-4c71-82fa-ecda02da3277","Type":"ContainerDied","Data":"6653bc0073fbdde40468fb06b953c1302846763f2888418b4fc21509ef3915d2"} Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.641238 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-2" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.641287 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"e367e739-45d9-4c71-82fa-ecda02da3277","Type":"ContainerDied","Data":"9165416a79a8d14934c00fc8e00a91ffd697d205964c3585f55278b965651da9"} Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.641311 4714 scope.go:117] "RemoveContainer" containerID="6653bc0073fbdde40468fb06b953c1302846763f2888418b4fc21509ef3915d2" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.646296 4714 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="cinder-kuttl-tests/keystone52f2-account-delete-qxsl4" secret="" err="secret \"galera-openstack-dockercfg-qdqvq\" not found" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.646338 4714 scope.go:117] "RemoveContainer" containerID="b8914d707e1c8c5545d8b139524bf6fc2f7931a0ae5843dc9cc4a838b4c891ba" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.645808 4714 generic.go:334] "Generic (PLEG): container finished" podID="112a1cde-5990-4140-97dd-c2bbc4f73197" containerID="b8914d707e1c8c5545d8b139524bf6fc2f7931a0ae5843dc9cc4a838b4c891ba" exitCode=1 Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.646864 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone52f2-account-delete-qxsl4" event={"ID":"112a1cde-5990-4140-97dd-c2bbc4f73197","Type":"ContainerDied","Data":"b8914d707e1c8c5545d8b139524bf6fc2f7931a0ae5843dc9cc4a838b4c891ba"} Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.646960 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/root-account-create-update-jg6sh" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.656419 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/root-account-create-update-jg6sh" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.671098 4714 scope.go:117] "RemoveContainer" containerID="ec556421e74871bfa9c85c53ff648a1a0691767027648a2904e427bc8f75360f" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.698036 4714 scope.go:117] "RemoveContainer" containerID="6653bc0073fbdde40468fb06b953c1302846763f2888418b4fc21509ef3915d2" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.698609 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e367e739-45d9-4c71-82fa-ecda02da3277-operator-scripts\") pod \"e367e739-45d9-4c71-82fa-ecda02da3277\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " Jan 29 16:30:37 crc kubenswrapper[4714]: E0129 16:30:37.698632 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6653bc0073fbdde40468fb06b953c1302846763f2888418b4fc21509ef3915d2\": container with ID starting with 6653bc0073fbdde40468fb06b953c1302846763f2888418b4fc21509ef3915d2 not found: ID does not exist" containerID="6653bc0073fbdde40468fb06b953c1302846763f2888418b4fc21509ef3915d2" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.698697 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6653bc0073fbdde40468fb06b953c1302846763f2888418b4fc21509ef3915d2"} err="failed to get container status \"6653bc0073fbdde40468fb06b953c1302846763f2888418b4fc21509ef3915d2\": rpc error: code = NotFound desc = could not find container \"6653bc0073fbdde40468fb06b953c1302846763f2888418b4fc21509ef3915d2\": container with ID starting with 6653bc0073fbdde40468fb06b953c1302846763f2888418b4fc21509ef3915d2 not found: ID does not exist" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.698648 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e367e739-45d9-4c71-82fa-ecda02da3277-kolla-config\") pod \"e367e739-45d9-4c71-82fa-ecda02da3277\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.698832 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"e367e739-45d9-4c71-82fa-ecda02da3277\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.699112 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e367e739-45d9-4c71-82fa-ecda02da3277-config-data-generated\") pod \"e367e739-45d9-4c71-82fa-ecda02da3277\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.699151 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nmlf\" (UniqueName: \"kubernetes.io/projected/e367e739-45d9-4c71-82fa-ecda02da3277-kube-api-access-8nmlf\") pod \"e367e739-45d9-4c71-82fa-ecda02da3277\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.699228 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e367e739-45d9-4c71-82fa-ecda02da3277-config-data-default\") pod \"e367e739-45d9-4c71-82fa-ecda02da3277\" (UID: \"e367e739-45d9-4c71-82fa-ecda02da3277\") " Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.699416 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e367e739-45d9-4c71-82fa-ecda02da3277-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "e367e739-45d9-4c71-82fa-ecda02da3277" (UID: "e367e739-45d9-4c71-82fa-ecda02da3277"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.699746 4714 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e367e739-45d9-4c71-82fa-ecda02da3277-kolla-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.699819 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e367e739-45d9-4c71-82fa-ecda02da3277-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e367e739-45d9-4c71-82fa-ecda02da3277" (UID: "e367e739-45d9-4c71-82fa-ecda02da3277"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.699841 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e367e739-45d9-4c71-82fa-ecda02da3277-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "e367e739-45d9-4c71-82fa-ecda02da3277" (UID: "e367e739-45d9-4c71-82fa-ecda02da3277"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:30:37 crc kubenswrapper[4714]: E0129 16:30:37.699919 4714 configmap.go:193] Couldn't get configMap cinder-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Jan 29 16:30:37 crc kubenswrapper[4714]: E0129 16:30:37.699997 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/112a1cde-5990-4140-97dd-c2bbc4f73197-operator-scripts podName:112a1cde-5990-4140-97dd-c2bbc4f73197 nodeName:}" failed. No retries permitted until 2026-01-29 16:30:38.199973388 +0000 UTC m=+1244.720474718 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/112a1cde-5990-4140-97dd-c2bbc4f73197-operator-scripts") pod "keystone52f2-account-delete-qxsl4" (UID: "112a1cde-5990-4140-97dd-c2bbc4f73197") : configmap "openstack-scripts" not found Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.700583 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e367e739-45d9-4c71-82fa-ecda02da3277-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "e367e739-45d9-4c71-82fa-ecda02da3277" (UID: "e367e739-45d9-4c71-82fa-ecda02da3277"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.700666 4714 scope.go:117] "RemoveContainer" containerID="ec556421e74871bfa9c85c53ff648a1a0691767027648a2904e427bc8f75360f" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.706136 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e367e739-45d9-4c71-82fa-ecda02da3277-kube-api-access-8nmlf" (OuterVolumeSpecName: "kube-api-access-8nmlf") pod "e367e739-45d9-4c71-82fa-ecda02da3277" (UID: "e367e739-45d9-4c71-82fa-ecda02da3277"). InnerVolumeSpecName "kube-api-access-8nmlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.713448 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "mysql-db") pod "e367e739-45d9-4c71-82fa-ecda02da3277" (UID: "e367e739-45d9-4c71-82fa-ecda02da3277"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 29 16:30:37 crc kubenswrapper[4714]: E0129 16:30:37.714633 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec556421e74871bfa9c85c53ff648a1a0691767027648a2904e427bc8f75360f\": container with ID starting with ec556421e74871bfa9c85c53ff648a1a0691767027648a2904e427bc8f75360f not found: ID does not exist" containerID="ec556421e74871bfa9c85c53ff648a1a0691767027648a2904e427bc8f75360f" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.714693 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec556421e74871bfa9c85c53ff648a1a0691767027648a2904e427bc8f75360f"} err="failed to get container status \"ec556421e74871bfa9c85c53ff648a1a0691767027648a2904e427bc8f75360f\": rpc error: code = NotFound desc = could not find container \"ec556421e74871bfa9c85c53ff648a1a0691767027648a2904e427bc8f75360f\": container with ID starting with ec556421e74871bfa9c85c53ff648a1a0691767027648a2904e427bc8f75360f not found: ID does not exist" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.801009 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e367e739-45d9-4c71-82fa-ecda02da3277-config-data-generated\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.801036 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nmlf\" (UniqueName: \"kubernetes.io/projected/e367e739-45d9-4c71-82fa-ecda02da3277-kube-api-access-8nmlf\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.801045 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e367e739-45d9-4c71-82fa-ecda02da3277-config-data-default\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.801057 4714 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e367e739-45d9-4c71-82fa-ecda02da3277-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.801090 4714 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.811685 4714 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.902152 4714 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.982001 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Jan 29 16:30:37 crc kubenswrapper[4714]: I0129 16:30:37.997069 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.001500 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.040344 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/rabbitmq-server-0" podUID="55e23ac1-a89b-4689-a17d-bee875f7783e" containerName="rabbitmq" containerID="cri-o://ee125ef7ce148229ad82b22d2a7578dd54b32a1f3c67998b7176056878a4676a" gracePeriod=604800 Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.191185 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c94e6f97-6224-46d2-b406-f5d02a596cb7" path="/var/lib/kubelet/pods/c94e6f97-6224-46d2-b406-f5d02a596cb7/volumes" Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.191719 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e367e739-45d9-4c71-82fa-ecda02da3277" path="/var/lib/kubelet/pods/e367e739-45d9-4c71-82fa-ecda02da3277/volumes" Jan 29 16:30:38 crc kubenswrapper[4714]: E0129 16:30:38.204775 4714 configmap.go:193] Couldn't get configMap cinder-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Jan 29 16:30:38 crc kubenswrapper[4714]: E0129 16:30:38.205040 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/112a1cde-5990-4140-97dd-c2bbc4f73197-operator-scripts podName:112a1cde-5990-4140-97dd-c2bbc4f73197 nodeName:}" failed. No retries permitted until 2026-01-29 16:30:39.205025766 +0000 UTC m=+1245.725526886 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/112a1cde-5990-4140-97dd-c2bbc4f73197-operator-scripts") pod "keystone52f2-account-delete-qxsl4" (UID: "112a1cde-5990-4140-97dd-c2bbc4f73197") : configmap "openstack-scripts" not found Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.307810 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc4cd\" (UniqueName: \"kubernetes.io/projected/4a2536dd-4258-4b5b-863c-a76431c992ee-kube-api-access-rc4cd\") pod \"root-account-create-update-jg6sh\" (UID: \"4a2536dd-4258-4b5b-863c-a76431c992ee\") " pod="cinder-kuttl-tests/root-account-create-update-jg6sh" Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.307852 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a2536dd-4258-4b5b-863c-a76431c992ee-operator-scripts\") pod \"root-account-create-update-jg6sh\" (UID: \"4a2536dd-4258-4b5b-863c-a76431c992ee\") " pod="cinder-kuttl-tests/root-account-create-update-jg6sh" Jan 29 16:30:38 crc kubenswrapper[4714]: E0129 16:30:38.309599 4714 configmap.go:193] Couldn't get configMap cinder-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Jan 29 16:30:38 crc kubenswrapper[4714]: E0129 16:30:38.309658 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4a2536dd-4258-4b5b-863c-a76431c992ee-operator-scripts podName:4a2536dd-4258-4b5b-863c-a76431c992ee nodeName:}" failed. No retries permitted until 2026-01-29 16:30:40.309645159 +0000 UTC m=+1246.830146279 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4a2536dd-4258-4b5b-863c-a76431c992ee-operator-scripts") pod "root-account-create-update-jg6sh" (UID: "4a2536dd-4258-4b5b-863c-a76431c992ee") : configmap "openstack-scripts" not found Jan 29 16:30:38 crc kubenswrapper[4714]: E0129 16:30:38.313206 4714 projected.go:194] Error preparing data for projected volume kube-api-access-rc4cd for pod cinder-kuttl-tests/root-account-create-update-jg6sh: failed to fetch token: serviceaccounts "galera-openstack" not found Jan 29 16:30:38 crc kubenswrapper[4714]: E0129 16:30:38.313271 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4a2536dd-4258-4b5b-863c-a76431c992ee-kube-api-access-rc4cd podName:4a2536dd-4258-4b5b-863c-a76431c992ee nodeName:}" failed. No retries permitted until 2026-01-29 16:30:40.313260833 +0000 UTC m=+1246.833761953 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-rc4cd" (UniqueName: "kubernetes.io/projected/4a2536dd-4258-4b5b-863c-a76431c992ee-kube-api-access-rc4cd") pod "root-account-create-update-jg6sh" (UID: "4a2536dd-4258-4b5b-863c-a76431c992ee") : failed to fetch token: serviceaccounts "galera-openstack" not found Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.458370 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/memcached-0" Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.509016 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea-config-data\") pod \"d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea\" (UID: \"d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea\") " Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.509092 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea-kolla-config\") pod \"d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea\" (UID: \"d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea\") " Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.509159 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkmzp\" (UniqueName: \"kubernetes.io/projected/d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea-kube-api-access-vkmzp\") pod \"d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea\" (UID: \"d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea\") " Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.509804 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea-config-data" (OuterVolumeSpecName: "config-data") pod "d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea" (UID: "d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.510295 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea" (UID: "d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.514207 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea-kube-api-access-vkmzp" (OuterVolumeSpecName: "kube-api-access-vkmzp") pod "d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea" (UID: "d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea"). InnerVolumeSpecName "kube-api-access-vkmzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.610587 4714 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea-kolla-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.610627 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkmzp\" (UniqueName: \"kubernetes.io/projected/d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea-kube-api-access-vkmzp\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.610640 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.658777 4714 generic.go:334] "Generic (PLEG): container finished" podID="d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea" containerID="5244ecc09d3c74f3a195f77af7759f9d242093032f09d5208383b4619dda295d" exitCode=0 Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.658886 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/memcached-0" Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.658879 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/memcached-0" event={"ID":"d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea","Type":"ContainerDied","Data":"5244ecc09d3c74f3a195f77af7759f9d242093032f09d5208383b4619dda295d"} Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.658998 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/memcached-0" event={"ID":"d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea","Type":"ContainerDied","Data":"5fcd1b55c77976e4d94c390473639a6b02e3c4a2129659d96d1e68f62ca74a39"} Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.659092 4714 scope.go:117] "RemoveContainer" containerID="5244ecc09d3c74f3a195f77af7759f9d242093032f09d5208383b4619dda295d" Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.662549 4714 generic.go:334] "Generic (PLEG): container finished" podID="112a1cde-5990-4140-97dd-c2bbc4f73197" containerID="cf816b8f59cc18e67ef8edcaed9119f0334f92970ee8c49b8887e2bc061f93ab" exitCode=1 Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.662642 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/root-account-create-update-jg6sh" Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.662735 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone52f2-account-delete-qxsl4" event={"ID":"112a1cde-5990-4140-97dd-c2bbc4f73197","Type":"ContainerDied","Data":"cf816b8f59cc18e67ef8edcaed9119f0334f92970ee8c49b8887e2bc061f93ab"} Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.663501 4714 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="cinder-kuttl-tests/keystone52f2-account-delete-qxsl4" secret="" err="secret \"galera-openstack-dockercfg-qdqvq\" not found" Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.663559 4714 scope.go:117] "RemoveContainer" containerID="cf816b8f59cc18e67ef8edcaed9119f0334f92970ee8c49b8887e2bc061f93ab" Jan 29 16:30:38 crc kubenswrapper[4714]: E0129 16:30:38.664047 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=keystone52f2-account-delete-qxsl4_cinder-kuttl-tests(112a1cde-5990-4140-97dd-c2bbc4f73197)\"" pod="cinder-kuttl-tests/keystone52f2-account-delete-qxsl4" podUID="112a1cde-5990-4140-97dd-c2bbc4f73197" Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.723332 4714 scope.go:117] "RemoveContainer" containerID="5244ecc09d3c74f3a195f77af7759f9d242093032f09d5208383b4619dda295d" Jan 29 16:30:38 crc kubenswrapper[4714]: E0129 16:30:38.727092 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5244ecc09d3c74f3a195f77af7759f9d242093032f09d5208383b4619dda295d\": container with ID starting with 5244ecc09d3c74f3a195f77af7759f9d242093032f09d5208383b4619dda295d not found: ID does not exist" containerID="5244ecc09d3c74f3a195f77af7759f9d242093032f09d5208383b4619dda295d" Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.727136 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5244ecc09d3c74f3a195f77af7759f9d242093032f09d5208383b4619dda295d"} err="failed to get container status \"5244ecc09d3c74f3a195f77af7759f9d242093032f09d5208383b4619dda295d\": rpc error: code = NotFound desc = could not find container \"5244ecc09d3c74f3a195f77af7759f9d242093032f09d5208383b4619dda295d\": container with ID starting with 5244ecc09d3c74f3a195f77af7759f9d242093032f09d5208383b4619dda295d not found: ID does not exist" Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.727193 4714 scope.go:117] "RemoveContainer" containerID="b8914d707e1c8c5545d8b139524bf6fc2f7931a0ae5843dc9cc4a838b4c891ba" Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.736026 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/root-account-create-update-jg6sh"] Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.761022 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/root-account-create-update-jg6sh"] Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.798075 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.800072 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.819957 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc4cd\" (UniqueName: \"kubernetes.io/projected/4a2536dd-4258-4b5b-863c-a76431c992ee-kube-api-access-rc4cd\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.819998 4714 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a2536dd-4258-4b5b-863c-a76431c992ee-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:38 crc kubenswrapper[4714]: I0129 16:30:38.859555 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/openstack-galera-1" podUID="f8d336f2-b190-4e32-be3a-27fbf0e50a06" containerName="galera" containerID="cri-o://83c3eb1ecb12cd8202e4a2f2b14330aa7199092be01cfe200e08827657c44a8b" gracePeriod=28 Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.006654 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs"] Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.006915 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" podUID="9aa31790-7a3c-4a66-aace-c087c0221c6b" containerName="manager" containerID="cri-o://36513d8368599e830d1d92fc041ab07bd030a4ddf286f084b21d04377e2a5dbd" gracePeriod=10 Jan 29 16:30:39 crc kubenswrapper[4714]: E0129 16:30:39.227601 4714 configmap.go:193] Couldn't get configMap cinder-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Jan 29 16:30:39 crc kubenswrapper[4714]: E0129 16:30:39.227668 4714 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/112a1cde-5990-4140-97dd-c2bbc4f73197-operator-scripts podName:112a1cde-5990-4140-97dd-c2bbc4f73197 nodeName:}" failed. No retries permitted until 2026-01-29 16:30:41.227649492 +0000 UTC m=+1247.748150612 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/112a1cde-5990-4140-97dd-c2bbc4f73197-operator-scripts") pod "keystone52f2-account-delete-qxsl4" (UID: "112a1cde-5990-4140-97dd-c2bbc4f73197") : configmap "openstack-scripts" not found Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.333276 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/cinder-operator-index-d7f6m"] Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.333818 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/cinder-operator-index-d7f6m" podUID="87506df3-b56a-4598-8309-e865dc93cf53" containerName="registry-server" containerID="cri-o://ddec768853957c5a9f807fefab5f4e26d2050966765bc9353326a1334ad4a549" gracePeriod=30 Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.387390 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6"] Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.394683 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/60e6154c28b7f915d22d12701456edcde02ff17009fbb6fad32c7757df6b2j6"] Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.439529 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.544645 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.547383 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-credential-keys\") pod \"fc06a535-6f60-438e-b52d-5dc90fae8c67\" (UID: \"fc06a535-6f60-438e-b52d-5dc90fae8c67\") " Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.547432 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-fernet-keys\") pod \"fc06a535-6f60-438e-b52d-5dc90fae8c67\" (UID: \"fc06a535-6f60-438e-b52d-5dc90fae8c67\") " Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.547580 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-scripts\") pod \"fc06a535-6f60-438e-b52d-5dc90fae8c67\" (UID: \"fc06a535-6f60-438e-b52d-5dc90fae8c67\") " Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.547629 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-config-data\") pod \"fc06a535-6f60-438e-b52d-5dc90fae8c67\" (UID: \"fc06a535-6f60-438e-b52d-5dc90fae8c67\") " Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.547656 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2vn9\" (UniqueName: \"kubernetes.io/projected/fc06a535-6f60-438e-b52d-5dc90fae8c67-kube-api-access-c2vn9\") pod \"fc06a535-6f60-438e-b52d-5dc90fae8c67\" (UID: \"fc06a535-6f60-438e-b52d-5dc90fae8c67\") " Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.552348 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc06a535-6f60-438e-b52d-5dc90fae8c67-kube-api-access-c2vn9" (OuterVolumeSpecName: "kube-api-access-c2vn9") pod "fc06a535-6f60-438e-b52d-5dc90fae8c67" (UID: "fc06a535-6f60-438e-b52d-5dc90fae8c67"). InnerVolumeSpecName "kube-api-access-c2vn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.552561 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "fc06a535-6f60-438e-b52d-5dc90fae8c67" (UID: "fc06a535-6f60-438e-b52d-5dc90fae8c67"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.554118 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-scripts" (OuterVolumeSpecName: "scripts") pod "fc06a535-6f60-438e-b52d-5dc90fae8c67" (UID: "fc06a535-6f60-438e-b52d-5dc90fae8c67"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.568143 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "fc06a535-6f60-438e-b52d-5dc90fae8c67" (UID: "fc06a535-6f60-438e-b52d-5dc90fae8c67"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:39 crc kubenswrapper[4714]: E0129 16:30:39.568192 4714 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ddec768853957c5a9f807fefab5f4e26d2050966765bc9353326a1334ad4a549 is running failed: container process not found" containerID="ddec768853957c5a9f807fefab5f4e26d2050966765bc9353326a1334ad4a549" cmd=["grpc_health_probe","-addr=:50051"] Jan 29 16:30:39 crc kubenswrapper[4714]: E0129 16:30:39.569285 4714 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ddec768853957c5a9f807fefab5f4e26d2050966765bc9353326a1334ad4a549 is running failed: container process not found" containerID="ddec768853957c5a9f807fefab5f4e26d2050966765bc9353326a1334ad4a549" cmd=["grpc_health_probe","-addr=:50051"] Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.576952 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-config-data" (OuterVolumeSpecName: "config-data") pod "fc06a535-6f60-438e-b52d-5dc90fae8c67" (UID: "fc06a535-6f60-438e-b52d-5dc90fae8c67"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:39 crc kubenswrapper[4714]: E0129 16:30:39.578464 4714 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ddec768853957c5a9f807fefab5f4e26d2050966765bc9353326a1334ad4a549 is running failed: container process not found" containerID="ddec768853957c5a9f807fefab5f4e26d2050966765bc9353326a1334ad4a549" cmd=["grpc_health_probe","-addr=:50051"] Jan 29 16:30:39 crc kubenswrapper[4714]: E0129 16:30:39.578511 4714 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ddec768853957c5a9f807fefab5f4e26d2050966765bc9353326a1334ad4a549 is running failed: container process not found" probeType="Readiness" pod="openstack-operators/cinder-operator-index-d7f6m" podUID="87506df3-b56a-4598-8309-e865dc93cf53" containerName="registry-server" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.646072 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.655527 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4dvn\" (UniqueName: \"kubernetes.io/projected/9aa31790-7a3c-4a66-aace-c087c0221c6b-kube-api-access-t4dvn\") pod \"9aa31790-7a3c-4a66-aace-c087c0221c6b\" (UID: \"9aa31790-7a3c-4a66-aace-c087c0221c6b\") " Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.655594 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9aa31790-7a3c-4a66-aace-c087c0221c6b-apiservice-cert\") pod \"9aa31790-7a3c-4a66-aace-c087c0221c6b\" (UID: \"9aa31790-7a3c-4a66-aace-c087c0221c6b\") " Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.655661 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9aa31790-7a3c-4a66-aace-c087c0221c6b-webhook-cert\") pod \"9aa31790-7a3c-4a66-aace-c087c0221c6b\" (UID: \"9aa31790-7a3c-4a66-aace-c087c0221c6b\") " Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.655987 4714 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.656008 4714 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-config-data\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.656022 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2vn9\" (UniqueName: \"kubernetes.io/projected/fc06a535-6f60-438e-b52d-5dc90fae8c67-kube-api-access-c2vn9\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.656036 4714 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.656047 4714 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc06a535-6f60-438e-b52d-5dc90fae8c67-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.663829 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa31790-7a3c-4a66-aace-c087c0221c6b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "9aa31790-7a3c-4a66-aace-c087c0221c6b" (UID: "9aa31790-7a3c-4a66-aace-c087c0221c6b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.674192 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa31790-7a3c-4a66-aace-c087c0221c6b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "9aa31790-7a3c-4a66-aace-c087c0221c6b" (UID: "9aa31790-7a3c-4a66-aace-c087c0221c6b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.674241 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aa31790-7a3c-4a66-aace-c087c0221c6b-kube-api-access-t4dvn" (OuterVolumeSpecName: "kube-api-access-t4dvn") pod "9aa31790-7a3c-4a66-aace-c087c0221c6b" (UID: "9aa31790-7a3c-4a66-aace-c087c0221c6b"). InnerVolumeSpecName "kube-api-access-t4dvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.692916 4714 generic.go:334] "Generic (PLEG): container finished" podID="87506df3-b56a-4598-8309-e865dc93cf53" containerID="ddec768853957c5a9f807fefab5f4e26d2050966765bc9353326a1334ad4a549" exitCode=0 Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.692991 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-index-d7f6m" event={"ID":"87506df3-b56a-4598-8309-e865dc93cf53","Type":"ContainerDied","Data":"ddec768853957c5a9f807fefab5f4e26d2050966765bc9353326a1334ad4a549"} Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.696861 4714 generic.go:334] "Generic (PLEG): container finished" podID="fc06a535-6f60-438e-b52d-5dc90fae8c67" containerID="4adeb441736bb7d27c549844b9147979d27f5dbed7e42a9b990619d275a00000" exitCode=0 Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.696947 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" event={"ID":"fc06a535-6f60-438e-b52d-5dc90fae8c67","Type":"ContainerDied","Data":"4adeb441736bb7d27c549844b9147979d27f5dbed7e42a9b990619d275a00000"} Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.696978 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" event={"ID":"fc06a535-6f60-438e-b52d-5dc90fae8c67","Type":"ContainerDied","Data":"a35aebf0427b3f34153f6b20222d8016725e500d95a3072063dc0d02bd8d902e"} Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.697000 4714 scope.go:117] "RemoveContainer" containerID="4adeb441736bb7d27c549844b9147979d27f5dbed7e42a9b990619d275a00000" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.697112 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db9b49999-6gd95" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.710355 4714 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="cinder-kuttl-tests/keystone52f2-account-delete-qxsl4" secret="" err="secret \"galera-openstack-dockercfg-qdqvq\" not found" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.711986 4714 scope.go:117] "RemoveContainer" containerID="cf816b8f59cc18e67ef8edcaed9119f0334f92970ee8c49b8887e2bc061f93ab" Jan 29 16:30:39 crc kubenswrapper[4714]: E0129 16:30:39.712676 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=keystone52f2-account-delete-qxsl4_cinder-kuttl-tests(112a1cde-5990-4140-97dd-c2bbc4f73197)\"" pod="cinder-kuttl-tests/keystone52f2-account-delete-qxsl4" podUID="112a1cde-5990-4140-97dd-c2bbc4f73197" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.713650 4714 generic.go:334] "Generic (PLEG): container finished" podID="55e23ac1-a89b-4689-a17d-bee875f7783e" containerID="ee125ef7ce148229ad82b22d2a7578dd54b32a1f3c67998b7176056878a4676a" exitCode=0 Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.713693 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/rabbitmq-server-0" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.713766 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"55e23ac1-a89b-4689-a17d-bee875f7783e","Type":"ContainerDied","Data":"ee125ef7ce148229ad82b22d2a7578dd54b32a1f3c67998b7176056878a4676a"} Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.713844 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"55e23ac1-a89b-4689-a17d-bee875f7783e","Type":"ContainerDied","Data":"299eb5904909cd50412aad25a871f9888d290b3d4f1acfe53103c96e6f05a1bc"} Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.715298 4714 generic.go:334] "Generic (PLEG): container finished" podID="9aa31790-7a3c-4a66-aace-c087c0221c6b" containerID="36513d8368599e830d1d92fc041ab07bd030a4ddf286f084b21d04377e2a5dbd" exitCode=0 Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.715487 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" event={"ID":"9aa31790-7a3c-4a66-aace-c087c0221c6b","Type":"ContainerDied","Data":"36513d8368599e830d1d92fc041ab07bd030a4ddf286f084b21d04377e2a5dbd"} Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.715556 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" event={"ID":"9aa31790-7a3c-4a66-aace-c087c0221c6b","Type":"ContainerDied","Data":"c82d440d4556664aec8a776ead06b3d925538dac3151f1c3b85d4cf089d48d43"} Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.715640 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.727873 4714 scope.go:117] "RemoveContainer" containerID="4adeb441736bb7d27c549844b9147979d27f5dbed7e42a9b990619d275a00000" Jan 29 16:30:39 crc kubenswrapper[4714]: E0129 16:30:39.728465 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4adeb441736bb7d27c549844b9147979d27f5dbed7e42a9b990619d275a00000\": container with ID starting with 4adeb441736bb7d27c549844b9147979d27f5dbed7e42a9b990619d275a00000 not found: ID does not exist" containerID="4adeb441736bb7d27c549844b9147979d27f5dbed7e42a9b990619d275a00000" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.728499 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4adeb441736bb7d27c549844b9147979d27f5dbed7e42a9b990619d275a00000"} err="failed to get container status \"4adeb441736bb7d27c549844b9147979d27f5dbed7e42a9b990619d275a00000\": rpc error: code = NotFound desc = could not find container \"4adeb441736bb7d27c549844b9147979d27f5dbed7e42a9b990619d275a00000\": container with ID starting with 4adeb441736bb7d27c549844b9147979d27f5dbed7e42a9b990619d275a00000 not found: ID does not exist" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.728544 4714 scope.go:117] "RemoveContainer" containerID="ee125ef7ce148229ad82b22d2a7578dd54b32a1f3c67998b7176056878a4676a" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.744839 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-db9b49999-6gd95"] Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.750610 4714 scope.go:117] "RemoveContainer" containerID="eb95fbc4965d0aeffea6fce3a32742ee3a1bc5446fd33b5341ed6d4be042d493" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.756009 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-db9b49999-6gd95"] Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.758519 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-21c03d9a-bbed-44e7-9b38-49f8915c54d2\") pod \"55e23ac1-a89b-4689-a17d-bee875f7783e\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.758631 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pgjf\" (UniqueName: \"kubernetes.io/projected/55e23ac1-a89b-4689-a17d-bee875f7783e-kube-api-access-8pgjf\") pod \"55e23ac1-a89b-4689-a17d-bee875f7783e\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.758707 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/55e23ac1-a89b-4689-a17d-bee875f7783e-plugins-conf\") pod \"55e23ac1-a89b-4689-a17d-bee875f7783e\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.758804 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/55e23ac1-a89b-4689-a17d-bee875f7783e-rabbitmq-confd\") pod \"55e23ac1-a89b-4689-a17d-bee875f7783e\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.758883 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/55e23ac1-a89b-4689-a17d-bee875f7783e-rabbitmq-erlang-cookie\") pod \"55e23ac1-a89b-4689-a17d-bee875f7783e\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.758975 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/55e23ac1-a89b-4689-a17d-bee875f7783e-rabbitmq-plugins\") pod \"55e23ac1-a89b-4689-a17d-bee875f7783e\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.759826 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/55e23ac1-a89b-4689-a17d-bee875f7783e-erlang-cookie-secret\") pod \"55e23ac1-a89b-4689-a17d-bee875f7783e\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.759968 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/55e23ac1-a89b-4689-a17d-bee875f7783e-pod-info\") pod \"55e23ac1-a89b-4689-a17d-bee875f7783e\" (UID: \"55e23ac1-a89b-4689-a17d-bee875f7783e\") " Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.759611 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55e23ac1-a89b-4689-a17d-bee875f7783e-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "55e23ac1-a89b-4689-a17d-bee875f7783e" (UID: "55e23ac1-a89b-4689-a17d-bee875f7783e"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.759642 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55e23ac1-a89b-4689-a17d-bee875f7783e-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "55e23ac1-a89b-4689-a17d-bee875f7783e" (UID: "55e23ac1-a89b-4689-a17d-bee875f7783e"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.759818 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55e23ac1-a89b-4689-a17d-bee875f7783e-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "55e23ac1-a89b-4689-a17d-bee875f7783e" (UID: "55e23ac1-a89b-4689-a17d-bee875f7783e"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.760839 4714 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9aa31790-7a3c-4a66-aace-c087c0221c6b-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.761173 4714 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/55e23ac1-a89b-4689-a17d-bee875f7783e-plugins-conf\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.761227 4714 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/55e23ac1-a89b-4689-a17d-bee875f7783e-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.761350 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4dvn\" (UniqueName: \"kubernetes.io/projected/9aa31790-7a3c-4a66-aace-c087c0221c6b-kube-api-access-t4dvn\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.761419 4714 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/55e23ac1-a89b-4689-a17d-bee875f7783e-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.761494 4714 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9aa31790-7a3c-4a66-aace-c087c0221c6b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.761822 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55e23ac1-a89b-4689-a17d-bee875f7783e-kube-api-access-8pgjf" (OuterVolumeSpecName: "kube-api-access-8pgjf") pod "55e23ac1-a89b-4689-a17d-bee875f7783e" (UID: "55e23ac1-a89b-4689-a17d-bee875f7783e"). InnerVolumeSpecName "kube-api-access-8pgjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.762180 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs"] Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.763574 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55e23ac1-a89b-4689-a17d-bee875f7783e-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "55e23ac1-a89b-4689-a17d-bee875f7783e" (UID: "55e23ac1-a89b-4689-a17d-bee875f7783e"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.764243 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/55e23ac1-a89b-4689-a17d-bee875f7783e-pod-info" (OuterVolumeSpecName: "pod-info") pod "55e23ac1-a89b-4689-a17d-bee875f7783e" (UID: "55e23ac1-a89b-4689-a17d-bee875f7783e"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.767690 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5fc6d4b6f5-9mdcs"] Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.776871 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-21c03d9a-bbed-44e7-9b38-49f8915c54d2" (OuterVolumeSpecName: "persistence") pod "55e23ac1-a89b-4689-a17d-bee875f7783e" (UID: "55e23ac1-a89b-4689-a17d-bee875f7783e"). InnerVolumeSpecName "pvc-21c03d9a-bbed-44e7-9b38-49f8915c54d2". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.777803 4714 scope.go:117] "RemoveContainer" containerID="ee125ef7ce148229ad82b22d2a7578dd54b32a1f3c67998b7176056878a4676a" Jan 29 16:30:39 crc kubenswrapper[4714]: E0129 16:30:39.779221 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee125ef7ce148229ad82b22d2a7578dd54b32a1f3c67998b7176056878a4676a\": container with ID starting with ee125ef7ce148229ad82b22d2a7578dd54b32a1f3c67998b7176056878a4676a not found: ID does not exist" containerID="ee125ef7ce148229ad82b22d2a7578dd54b32a1f3c67998b7176056878a4676a" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.779279 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee125ef7ce148229ad82b22d2a7578dd54b32a1f3c67998b7176056878a4676a"} err="failed to get container status \"ee125ef7ce148229ad82b22d2a7578dd54b32a1f3c67998b7176056878a4676a\": rpc error: code = NotFound desc = could not find container \"ee125ef7ce148229ad82b22d2a7578dd54b32a1f3c67998b7176056878a4676a\": container with ID starting with ee125ef7ce148229ad82b22d2a7578dd54b32a1f3c67998b7176056878a4676a not found: ID does not exist" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.779305 4714 scope.go:117] "RemoveContainer" containerID="eb95fbc4965d0aeffea6fce3a32742ee3a1bc5446fd33b5341ed6d4be042d493" Jan 29 16:30:39 crc kubenswrapper[4714]: E0129 16:30:39.779561 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb95fbc4965d0aeffea6fce3a32742ee3a1bc5446fd33b5341ed6d4be042d493\": container with ID starting with eb95fbc4965d0aeffea6fce3a32742ee3a1bc5446fd33b5341ed6d4be042d493 not found: ID does not exist" containerID="eb95fbc4965d0aeffea6fce3a32742ee3a1bc5446fd33b5341ed6d4be042d493" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.779606 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb95fbc4965d0aeffea6fce3a32742ee3a1bc5446fd33b5341ed6d4be042d493"} err="failed to get container status \"eb95fbc4965d0aeffea6fce3a32742ee3a1bc5446fd33b5341ed6d4be042d493\": rpc error: code = NotFound desc = could not find container \"eb95fbc4965d0aeffea6fce3a32742ee3a1bc5446fd33b5341ed6d4be042d493\": container with ID starting with eb95fbc4965d0aeffea6fce3a32742ee3a1bc5446fd33b5341ed6d4be042d493 not found: ID does not exist" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.779621 4714 scope.go:117] "RemoveContainer" containerID="36513d8368599e830d1d92fc041ab07bd030a4ddf286f084b21d04377e2a5dbd" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.779867 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-index-d7f6m" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.802623 4714 scope.go:117] "RemoveContainer" containerID="36513d8368599e830d1d92fc041ab07bd030a4ddf286f084b21d04377e2a5dbd" Jan 29 16:30:39 crc kubenswrapper[4714]: E0129 16:30:39.803118 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36513d8368599e830d1d92fc041ab07bd030a4ddf286f084b21d04377e2a5dbd\": container with ID starting with 36513d8368599e830d1d92fc041ab07bd030a4ddf286f084b21d04377e2a5dbd not found: ID does not exist" containerID="36513d8368599e830d1d92fc041ab07bd030a4ddf286f084b21d04377e2a5dbd" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.803179 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36513d8368599e830d1d92fc041ab07bd030a4ddf286f084b21d04377e2a5dbd"} err="failed to get container status \"36513d8368599e830d1d92fc041ab07bd030a4ddf286f084b21d04377e2a5dbd\": rpc error: code = NotFound desc = could not find container \"36513d8368599e830d1d92fc041ab07bd030a4ddf286f084b21d04377e2a5dbd\": container with ID starting with 36513d8368599e830d1d92fc041ab07bd030a4ddf286f084b21d04377e2a5dbd not found: ID does not exist" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.831611 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55e23ac1-a89b-4689-a17d-bee875f7783e-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "55e23ac1-a89b-4689-a17d-bee875f7783e" (UID: "55e23ac1-a89b-4689-a17d-bee875f7783e"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.862837 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6lgj\" (UniqueName: \"kubernetes.io/projected/87506df3-b56a-4598-8309-e865dc93cf53-kube-api-access-m6lgj\") pod \"87506df3-b56a-4598-8309-e865dc93cf53\" (UID: \"87506df3-b56a-4598-8309-e865dc93cf53\") " Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.863234 4714 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/55e23ac1-a89b-4689-a17d-bee875f7783e-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.863276 4714 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/55e23ac1-a89b-4689-a17d-bee875f7783e-pod-info\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.863315 4714 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-21c03d9a-bbed-44e7-9b38-49f8915c54d2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-21c03d9a-bbed-44e7-9b38-49f8915c54d2\") on node \"crc\" " Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.863330 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pgjf\" (UniqueName: \"kubernetes.io/projected/55e23ac1-a89b-4689-a17d-bee875f7783e-kube-api-access-8pgjf\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.863367 4714 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/55e23ac1-a89b-4689-a17d-bee875f7783e-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.865791 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87506df3-b56a-4598-8309-e865dc93cf53-kube-api-access-m6lgj" (OuterVolumeSpecName: "kube-api-access-m6lgj") pod "87506df3-b56a-4598-8309-e865dc93cf53" (UID: "87506df3-b56a-4598-8309-e865dc93cf53"). InnerVolumeSpecName "kube-api-access-m6lgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.882821 4714 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.883041 4714 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-21c03d9a-bbed-44e7-9b38-49f8915c54d2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-21c03d9a-bbed-44e7-9b38-49f8915c54d2") on node "crc" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.965035 4714 reconciler_common.go:293] "Volume detached for volume \"pvc-21c03d9a-bbed-44e7-9b38-49f8915c54d2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-21c03d9a-bbed-44e7-9b38-49f8915c54d2\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:39 crc kubenswrapper[4714]: I0129 16:30:39.965091 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6lgj\" (UniqueName: \"kubernetes.io/projected/87506df3-b56a-4598-8309-e865dc93cf53-kube-api-access-m6lgj\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.044375 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.048475 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.191598 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0eecc358-9581-489e-97ae-f600d35a7613" path="/var/lib/kubelet/pods/0eecc358-9581-489e-97ae-f600d35a7613/volumes" Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.192428 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a2536dd-4258-4b5b-863c-a76431c992ee" path="/var/lib/kubelet/pods/4a2536dd-4258-4b5b-863c-a76431c992ee/volumes" Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.192959 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55e23ac1-a89b-4689-a17d-bee875f7783e" path="/var/lib/kubelet/pods/55e23ac1-a89b-4689-a17d-bee875f7783e/volumes" Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.194073 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aa31790-7a3c-4a66-aace-c087c0221c6b" path="/var/lib/kubelet/pods/9aa31790-7a3c-4a66-aace-c087c0221c6b/volumes" Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.194833 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea" path="/var/lib/kubelet/pods/d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea/volumes" Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.195366 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc06a535-6f60-438e-b52d-5dc90fae8c67" path="/var/lib/kubelet/pods/fc06a535-6f60-438e-b52d-5dc90fae8c67/volumes" Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.726148 4714 generic.go:334] "Generic (PLEG): container finished" podID="f8d336f2-b190-4e32-be3a-27fbf0e50a06" containerID="83c3eb1ecb12cd8202e4a2f2b14330aa7199092be01cfe200e08827657c44a8b" exitCode=0 Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.726239 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"f8d336f2-b190-4e32-be3a-27fbf0e50a06","Type":"ContainerDied","Data":"83c3eb1ecb12cd8202e4a2f2b14330aa7199092be01cfe200e08827657c44a8b"} Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.728291 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-index-d7f6m" event={"ID":"87506df3-b56a-4598-8309-e865dc93cf53","Type":"ContainerDied","Data":"6f87b469ba7b044e8c1285048b5bce1bfd385e6eeae34b8f824127e313741cf2"} Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.728341 4714 scope.go:117] "RemoveContainer" containerID="ddec768853957c5a9f807fefab5f4e26d2050966765bc9353326a1334ad4a549" Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.728434 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-index-d7f6m" Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.767033 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/cinder-operator-index-d7f6m"] Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.769731 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/cinder-operator-index-d7f6m"] Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.800035 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-db-create-4hhfq"] Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.809611 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-db-create-4hhfq"] Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.819091 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-52f2-account-create-update-5x6mg"] Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.822854 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone52f2-account-delete-qxsl4"] Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.828994 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-52f2-account-create-update-5x6mg"] Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.868141 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.883947 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/openstack-galera-0" podUID="e27f02c1-a7d5-4d49-838b-df5445720a07" containerName="galera" containerID="cri-o://b027a08741bc7af4590803c43740e388681a5f71be01b0ea2974650f0e3e74fc" gracePeriod=26 Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.981307 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f8d336f2-b190-4e32-be3a-27fbf0e50a06-config-data-generated\") pod \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.981419 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f8d336f2-b190-4e32-be3a-27fbf0e50a06-config-data-default\") pod \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.981487 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f8d336f2-b190-4e32-be3a-27fbf0e50a06-kolla-config\") pod \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.981551 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f8d336f2-b190-4e32-be3a-27fbf0e50a06-operator-scripts\") pod \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.981580 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.981726 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhbgj\" (UniqueName: \"kubernetes.io/projected/f8d336f2-b190-4e32-be3a-27fbf0e50a06-kube-api-access-lhbgj\") pod \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\" (UID: \"f8d336f2-b190-4e32-be3a-27fbf0e50a06\") " Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.982742 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8d336f2-b190-4e32-be3a-27fbf0e50a06-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "f8d336f2-b190-4e32-be3a-27fbf0e50a06" (UID: "f8d336f2-b190-4e32-be3a-27fbf0e50a06"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.983435 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8d336f2-b190-4e32-be3a-27fbf0e50a06-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "f8d336f2-b190-4e32-be3a-27fbf0e50a06" (UID: "f8d336f2-b190-4e32-be3a-27fbf0e50a06"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.983694 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8d336f2-b190-4e32-be3a-27fbf0e50a06-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f8d336f2-b190-4e32-be3a-27fbf0e50a06" (UID: "f8d336f2-b190-4e32-be3a-27fbf0e50a06"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.983987 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8d336f2-b190-4e32-be3a-27fbf0e50a06-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "f8d336f2-b190-4e32-be3a-27fbf0e50a06" (UID: "f8d336f2-b190-4e32-be3a-27fbf0e50a06"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:30:40 crc kubenswrapper[4714]: I0129 16:30:40.995792 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8d336f2-b190-4e32-be3a-27fbf0e50a06-kube-api-access-lhbgj" (OuterVolumeSpecName: "kube-api-access-lhbgj") pod "f8d336f2-b190-4e32-be3a-27fbf0e50a06" (UID: "f8d336f2-b190-4e32-be3a-27fbf0e50a06"). InnerVolumeSpecName "kube-api-access-lhbgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.002339 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "mysql-db") pod "f8d336f2-b190-4e32-be3a-27fbf0e50a06" (UID: "f8d336f2-b190-4e32-be3a-27fbf0e50a06"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.083466 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f8d336f2-b190-4e32-be3a-27fbf0e50a06-config-data-default\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.083830 4714 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f8d336f2-b190-4e32-be3a-27fbf0e50a06-kolla-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.083845 4714 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f8d336f2-b190-4e32-be3a-27fbf0e50a06-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.083873 4714 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.083886 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhbgj\" (UniqueName: \"kubernetes.io/projected/f8d336f2-b190-4e32-be3a-27fbf0e50a06-kube-api-access-lhbgj\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.083900 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f8d336f2-b190-4e32-be3a-27fbf0e50a06-config-data-generated\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.090512 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone52f2-account-delete-qxsl4" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.098923 4714 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.184442 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrfh5\" (UniqueName: \"kubernetes.io/projected/112a1cde-5990-4140-97dd-c2bbc4f73197-kube-api-access-rrfh5\") pod \"112a1cde-5990-4140-97dd-c2bbc4f73197\" (UID: \"112a1cde-5990-4140-97dd-c2bbc4f73197\") " Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.184546 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/112a1cde-5990-4140-97dd-c2bbc4f73197-operator-scripts\") pod \"112a1cde-5990-4140-97dd-c2bbc4f73197\" (UID: \"112a1cde-5990-4140-97dd-c2bbc4f73197\") " Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.184898 4714 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.185255 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/112a1cde-5990-4140-97dd-c2bbc4f73197-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "112a1cde-5990-4140-97dd-c2bbc4f73197" (UID: "112a1cde-5990-4140-97dd-c2bbc4f73197"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.189779 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/112a1cde-5990-4140-97dd-c2bbc4f73197-kube-api-access-rrfh5" (OuterVolumeSpecName: "kube-api-access-rrfh5") pod "112a1cde-5990-4140-97dd-c2bbc4f73197" (UID: "112a1cde-5990-4140-97dd-c2bbc4f73197"). InnerVolumeSpecName "kube-api-access-rrfh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.286497 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrfh5\" (UniqueName: \"kubernetes.io/projected/112a1cde-5990-4140-97dd-c2bbc4f73197-kube-api-access-rrfh5\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.286532 4714 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/112a1cde-5990-4140-97dd-c2bbc4f73197-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.585214 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.690429 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e27f02c1-a7d5-4d49-838b-df5445720a07-config-data-default\") pod \"e27f02c1-a7d5-4d49-838b-df5445720a07\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.690490 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6vvb\" (UniqueName: \"kubernetes.io/projected/e27f02c1-a7d5-4d49-838b-df5445720a07-kube-api-access-p6vvb\") pod \"e27f02c1-a7d5-4d49-838b-df5445720a07\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.690571 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e27f02c1-a7d5-4d49-838b-df5445720a07-operator-scripts\") pod \"e27f02c1-a7d5-4d49-838b-df5445720a07\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.690590 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e27f02c1-a7d5-4d49-838b-df5445720a07-kolla-config\") pod \"e27f02c1-a7d5-4d49-838b-df5445720a07\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.690616 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"e27f02c1-a7d5-4d49-838b-df5445720a07\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.690636 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e27f02c1-a7d5-4d49-838b-df5445720a07-config-data-generated\") pod \"e27f02c1-a7d5-4d49-838b-df5445720a07\" (UID: \"e27f02c1-a7d5-4d49-838b-df5445720a07\") " Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.691198 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e27f02c1-a7d5-4d49-838b-df5445720a07-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "e27f02c1-a7d5-4d49-838b-df5445720a07" (UID: "e27f02c1-a7d5-4d49-838b-df5445720a07"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.691211 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e27f02c1-a7d5-4d49-838b-df5445720a07-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "e27f02c1-a7d5-4d49-838b-df5445720a07" (UID: "e27f02c1-a7d5-4d49-838b-df5445720a07"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.691344 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e27f02c1-a7d5-4d49-838b-df5445720a07-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "e27f02c1-a7d5-4d49-838b-df5445720a07" (UID: "e27f02c1-a7d5-4d49-838b-df5445720a07"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.691453 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e27f02c1-a7d5-4d49-838b-df5445720a07-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e27f02c1-a7d5-4d49-838b-df5445720a07" (UID: "e27f02c1-a7d5-4d49-838b-df5445720a07"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.694569 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e27f02c1-a7d5-4d49-838b-df5445720a07-kube-api-access-p6vvb" (OuterVolumeSpecName: "kube-api-access-p6vvb") pod "e27f02c1-a7d5-4d49-838b-df5445720a07" (UID: "e27f02c1-a7d5-4d49-838b-df5445720a07"). InnerVolumeSpecName "kube-api-access-p6vvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.701199 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "mysql-db") pod "e27f02c1-a7d5-4d49-838b-df5445720a07" (UID: "e27f02c1-a7d5-4d49-838b-df5445720a07"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.747967 4714 generic.go:334] "Generic (PLEG): container finished" podID="e27f02c1-a7d5-4d49-838b-df5445720a07" containerID="b027a08741bc7af4590803c43740e388681a5f71be01b0ea2974650f0e3e74fc" exitCode=0 Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.748026 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-0" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.748041 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"e27f02c1-a7d5-4d49-838b-df5445720a07","Type":"ContainerDied","Data":"b027a08741bc7af4590803c43740e388681a5f71be01b0ea2974650f0e3e74fc"} Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.748293 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"e27f02c1-a7d5-4d49-838b-df5445720a07","Type":"ContainerDied","Data":"37a7ed7cc71d5ba4399190ff48b8e2d70a326be5e9ad5c8773900669dfc3740e"} Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.748320 4714 scope.go:117] "RemoveContainer" containerID="b027a08741bc7af4590803c43740e388681a5f71be01b0ea2974650f0e3e74fc" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.751184 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"f8d336f2-b190-4e32-be3a-27fbf0e50a06","Type":"ContainerDied","Data":"2730c5f20c68259dc37f00e3d986e43810a9d5fe85207a2c704b5656855e553b"} Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.751263 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-1" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.758511 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone52f2-account-delete-qxsl4" event={"ID":"112a1cde-5990-4140-97dd-c2bbc4f73197","Type":"ContainerDied","Data":"886af71ec9bb3cd0c3ff50ec50762de0bc190d913b4893fee8dbe7877dbfc19b"} Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.758552 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone52f2-account-delete-qxsl4" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.768468 4714 scope.go:117] "RemoveContainer" containerID="2c8a621ff05274e374040ac76aa490fcc2858782c45b93e547a0c03c210a4530" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.782162 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.786755 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.793113 4714 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e27f02c1-a7d5-4d49-838b-df5445720a07-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.793159 4714 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e27f02c1-a7d5-4d49-838b-df5445720a07-kolla-config\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.793203 4714 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.793223 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e27f02c1-a7d5-4d49-838b-df5445720a07-config-data-generated\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.793245 4714 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e27f02c1-a7d5-4d49-838b-df5445720a07-config-data-default\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.793264 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6vvb\" (UniqueName: \"kubernetes.io/projected/e27f02c1-a7d5-4d49-838b-df5445720a07-kube-api-access-p6vvb\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.794289 4714 scope.go:117] "RemoveContainer" containerID="b027a08741bc7af4590803c43740e388681a5f71be01b0ea2974650f0e3e74fc" Jan 29 16:30:41 crc kubenswrapper[4714]: E0129 16:30:41.795132 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b027a08741bc7af4590803c43740e388681a5f71be01b0ea2974650f0e3e74fc\": container with ID starting with b027a08741bc7af4590803c43740e388681a5f71be01b0ea2974650f0e3e74fc not found: ID does not exist" containerID="b027a08741bc7af4590803c43740e388681a5f71be01b0ea2974650f0e3e74fc" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.795162 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b027a08741bc7af4590803c43740e388681a5f71be01b0ea2974650f0e3e74fc"} err="failed to get container status \"b027a08741bc7af4590803c43740e388681a5f71be01b0ea2974650f0e3e74fc\": rpc error: code = NotFound desc = could not find container \"b027a08741bc7af4590803c43740e388681a5f71be01b0ea2974650f0e3e74fc\": container with ID starting with b027a08741bc7af4590803c43740e388681a5f71be01b0ea2974650f0e3e74fc not found: ID does not exist" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.795187 4714 scope.go:117] "RemoveContainer" containerID="2c8a621ff05274e374040ac76aa490fcc2858782c45b93e547a0c03c210a4530" Jan 29 16:30:41 crc kubenswrapper[4714]: E0129 16:30:41.797288 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c8a621ff05274e374040ac76aa490fcc2858782c45b93e547a0c03c210a4530\": container with ID starting with 2c8a621ff05274e374040ac76aa490fcc2858782c45b93e547a0c03c210a4530 not found: ID does not exist" containerID="2c8a621ff05274e374040ac76aa490fcc2858782c45b93e547a0c03c210a4530" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.797312 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c8a621ff05274e374040ac76aa490fcc2858782c45b93e547a0c03c210a4530"} err="failed to get container status \"2c8a621ff05274e374040ac76aa490fcc2858782c45b93e547a0c03c210a4530\": rpc error: code = NotFound desc = could not find container \"2c8a621ff05274e374040ac76aa490fcc2858782c45b93e547a0c03c210a4530\": container with ID starting with 2c8a621ff05274e374040ac76aa490fcc2858782c45b93e547a0c03c210a4530 not found: ID does not exist" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.797326 4714 scope.go:117] "RemoveContainer" containerID="83c3eb1ecb12cd8202e4a2f2b14330aa7199092be01cfe200e08827657c44a8b" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.803276 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone52f2-account-delete-qxsl4"] Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.812395 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone52f2-account-delete-qxsl4"] Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.815763 4714 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.818553 4714 scope.go:117] "RemoveContainer" containerID="7b624009e8962fd057296e2a9f997c5b0aab61c294b85ef5b94d41ebe8dd89e7" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.821343 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.828402 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.846265 4714 scope.go:117] "RemoveContainer" containerID="cf816b8f59cc18e67ef8edcaed9119f0334f92970ee8c49b8887e2bc061f93ab" Jan 29 16:30:41 crc kubenswrapper[4714]: I0129 16:30:41.894856 4714 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:42 crc kubenswrapper[4714]: I0129 16:30:42.192773 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="112a1cde-5990-4140-97dd-c2bbc4f73197" path="/var/lib/kubelet/pods/112a1cde-5990-4140-97dd-c2bbc4f73197/volumes" Jan 29 16:30:42 crc kubenswrapper[4714]: I0129 16:30:42.193432 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bf4c895-a323-452d-8329-cb69a752341c" path="/var/lib/kubelet/pods/4bf4c895-a323-452d-8329-cb69a752341c/volumes" Jan 29 16:30:42 crc kubenswrapper[4714]: I0129 16:30:42.193974 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87506df3-b56a-4598-8309-e865dc93cf53" path="/var/lib/kubelet/pods/87506df3-b56a-4598-8309-e865dc93cf53/volumes" Jan 29 16:30:42 crc kubenswrapper[4714]: I0129 16:30:42.195651 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e27f02c1-a7d5-4d49-838b-df5445720a07" path="/var/lib/kubelet/pods/e27f02c1-a7d5-4d49-838b-df5445720a07/volumes" Jan 29 16:30:42 crc kubenswrapper[4714]: I0129 16:30:42.196501 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6c31118-0f0d-46fb-a9fc-d135e234fe41" path="/var/lib/kubelet/pods/f6c31118-0f0d-46fb-a9fc-d135e234fe41/volumes" Jan 29 16:30:42 crc kubenswrapper[4714]: I0129 16:30:42.197351 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8d336f2-b190-4e32-be3a-27fbf0e50a06" path="/var/lib/kubelet/pods/f8d336f2-b190-4e32-be3a-27fbf0e50a06/volumes" Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.022785 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28"] Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.023635 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" podUID="5d602ee5-4171-4dc7-9852-88c6019696e1" containerName="manager" containerID="cri-o://0389615fc2e1a33814fa46fa1999a6f01822d1c7cd6bb84e89e87e7102d3acc5" gracePeriod=10 Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.266542 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-vtc5h"] Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.266755 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-index-vtc5h" podUID="668764e7-6295-4275-bcc9-24b680ec685f" containerName="registry-server" containerID="cri-o://dc243f10cd3a85e410a8968c999f4a42fc022a8b377403e54fe6779138d67aa8" gracePeriod=30 Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.301528 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg"] Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.306066 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqppkg"] Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.496826 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.631003 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7sdp\" (UniqueName: \"kubernetes.io/projected/5d602ee5-4171-4dc7-9852-88c6019696e1-kube-api-access-p7sdp\") pod \"5d602ee5-4171-4dc7-9852-88c6019696e1\" (UID: \"5d602ee5-4171-4dc7-9852-88c6019696e1\") " Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.631047 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5d602ee5-4171-4dc7-9852-88c6019696e1-apiservice-cert\") pod \"5d602ee5-4171-4dc7-9852-88c6019696e1\" (UID: \"5d602ee5-4171-4dc7-9852-88c6019696e1\") " Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.631068 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5d602ee5-4171-4dc7-9852-88c6019696e1-webhook-cert\") pod \"5d602ee5-4171-4dc7-9852-88c6019696e1\" (UID: \"5d602ee5-4171-4dc7-9852-88c6019696e1\") " Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.635488 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d602ee5-4171-4dc7-9852-88c6019696e1-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "5d602ee5-4171-4dc7-9852-88c6019696e1" (UID: "5d602ee5-4171-4dc7-9852-88c6019696e1"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.635508 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d602ee5-4171-4dc7-9852-88c6019696e1-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "5d602ee5-4171-4dc7-9852-88c6019696e1" (UID: "5d602ee5-4171-4dc7-9852-88c6019696e1"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.637046 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d602ee5-4171-4dc7-9852-88c6019696e1-kube-api-access-p7sdp" (OuterVolumeSpecName: "kube-api-access-p7sdp") pod "5d602ee5-4171-4dc7-9852-88c6019696e1" (UID: "5d602ee5-4171-4dc7-9852-88c6019696e1"). InnerVolumeSpecName "kube-api-access-p7sdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.683404 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-vtc5h" Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.733417 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsc27\" (UniqueName: \"kubernetes.io/projected/668764e7-6295-4275-bcc9-24b680ec685f-kube-api-access-wsc27\") pod \"668764e7-6295-4275-bcc9-24b680ec685f\" (UID: \"668764e7-6295-4275-bcc9-24b680ec685f\") " Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.733724 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7sdp\" (UniqueName: \"kubernetes.io/projected/5d602ee5-4171-4dc7-9852-88c6019696e1-kube-api-access-p7sdp\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.733742 4714 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5d602ee5-4171-4dc7-9852-88c6019696e1-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.733751 4714 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5d602ee5-4171-4dc7-9852-88c6019696e1-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.736372 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/668764e7-6295-4275-bcc9-24b680ec685f-kube-api-access-wsc27" (OuterVolumeSpecName: "kube-api-access-wsc27") pod "668764e7-6295-4275-bcc9-24b680ec685f" (UID: "668764e7-6295-4275-bcc9-24b680ec685f"). InnerVolumeSpecName "kube-api-access-wsc27". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.779510 4714 generic.go:334] "Generic (PLEG): container finished" podID="5d602ee5-4171-4dc7-9852-88c6019696e1" containerID="0389615fc2e1a33814fa46fa1999a6f01822d1c7cd6bb84e89e87e7102d3acc5" exitCode=0 Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.779568 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" event={"ID":"5d602ee5-4171-4dc7-9852-88c6019696e1","Type":"ContainerDied","Data":"0389615fc2e1a33814fa46fa1999a6f01822d1c7cd6bb84e89e87e7102d3acc5"} Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.779598 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" event={"ID":"5d602ee5-4171-4dc7-9852-88c6019696e1","Type":"ContainerDied","Data":"68e4148d363b3ff81741fb86667cb2d686f613c7db6f1f53e639c87301659f49"} Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.779617 4714 scope.go:117] "RemoveContainer" containerID="0389615fc2e1a33814fa46fa1999a6f01822d1c7cd6bb84e89e87e7102d3acc5" Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.779708 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28" Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.791246 4714 generic.go:334] "Generic (PLEG): container finished" podID="668764e7-6295-4275-bcc9-24b680ec685f" containerID="dc243f10cd3a85e410a8968c999f4a42fc022a8b377403e54fe6779138d67aa8" exitCode=0 Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.791276 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-vtc5h" Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.791282 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-vtc5h" event={"ID":"668764e7-6295-4275-bcc9-24b680ec685f","Type":"ContainerDied","Data":"dc243f10cd3a85e410a8968c999f4a42fc022a8b377403e54fe6779138d67aa8"} Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.791397 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-vtc5h" event={"ID":"668764e7-6295-4275-bcc9-24b680ec685f","Type":"ContainerDied","Data":"047c3adf9e0c960c54099b8f9a0a168467b6e29a2c3acff719ceb9bbe1f69c79"} Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.808598 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28"] Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.811213 4714 scope.go:117] "RemoveContainer" containerID="0389615fc2e1a33814fa46fa1999a6f01822d1c7cd6bb84e89e87e7102d3acc5" Jan 29 16:30:43 crc kubenswrapper[4714]: E0129 16:30:43.811605 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0389615fc2e1a33814fa46fa1999a6f01822d1c7cd6bb84e89e87e7102d3acc5\": container with ID starting with 0389615fc2e1a33814fa46fa1999a6f01822d1c7cd6bb84e89e87e7102d3acc5 not found: ID does not exist" containerID="0389615fc2e1a33814fa46fa1999a6f01822d1c7cd6bb84e89e87e7102d3acc5" Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.811637 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0389615fc2e1a33814fa46fa1999a6f01822d1c7cd6bb84e89e87e7102d3acc5"} err="failed to get container status \"0389615fc2e1a33814fa46fa1999a6f01822d1c7cd6bb84e89e87e7102d3acc5\": rpc error: code = NotFound desc = could not find container \"0389615fc2e1a33814fa46fa1999a6f01822d1c7cd6bb84e89e87e7102d3acc5\": container with ID starting with 0389615fc2e1a33814fa46fa1999a6f01822d1c7cd6bb84e89e87e7102d3acc5 not found: ID does not exist" Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.811658 4714 scope.go:117] "RemoveContainer" containerID="dc243f10cd3a85e410a8968c999f4a42fc022a8b377403e54fe6779138d67aa8" Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.814218 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b97656f4c-wwx28"] Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.825211 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-vtc5h"] Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.831054 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-index-vtc5h"] Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.833906 4714 scope.go:117] "RemoveContainer" containerID="dc243f10cd3a85e410a8968c999f4a42fc022a8b377403e54fe6779138d67aa8" Jan 29 16:30:43 crc kubenswrapper[4714]: E0129 16:30:43.834419 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc243f10cd3a85e410a8968c999f4a42fc022a8b377403e54fe6779138d67aa8\": container with ID starting with dc243f10cd3a85e410a8968c999f4a42fc022a8b377403e54fe6779138d67aa8 not found: ID does not exist" containerID="dc243f10cd3a85e410a8968c999f4a42fc022a8b377403e54fe6779138d67aa8" Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.834461 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc243f10cd3a85e410a8968c999f4a42fc022a8b377403e54fe6779138d67aa8"} err="failed to get container status \"dc243f10cd3a85e410a8968c999f4a42fc022a8b377403e54fe6779138d67aa8\": rpc error: code = NotFound desc = could not find container \"dc243f10cd3a85e410a8968c999f4a42fc022a8b377403e54fe6779138d67aa8\": container with ID starting with dc243f10cd3a85e410a8968c999f4a42fc022a8b377403e54fe6779138d67aa8 not found: ID does not exist" Jan 29 16:30:43 crc kubenswrapper[4714]: I0129 16:30:43.834547 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsc27\" (UniqueName: \"kubernetes.io/projected/668764e7-6295-4275-bcc9-24b680ec685f-kube-api-access-wsc27\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:44 crc kubenswrapper[4714]: I0129 16:30:44.192576 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d602ee5-4171-4dc7-9852-88c6019696e1" path="/var/lib/kubelet/pods/5d602ee5-4171-4dc7-9852-88c6019696e1/volumes" Jan 29 16:30:44 crc kubenswrapper[4714]: I0129 16:30:44.193154 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="668764e7-6295-4275-bcc9-24b680ec685f" path="/var/lib/kubelet/pods/668764e7-6295-4275-bcc9-24b680ec685f/volumes" Jan 29 16:30:44 crc kubenswrapper[4714]: I0129 16:30:44.193714 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd598a7a-34ba-4392-908b-c18d89648bb5" path="/var/lib/kubelet/pods/dd598a7a-34ba-4392-908b-c18d89648bb5/volumes" Jan 29 16:30:45 crc kubenswrapper[4714]: I0129 16:30:45.838038 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-xprqx"] Jan 29 16:30:45 crc kubenswrapper[4714]: I0129 16:30:45.838914 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-xprqx" podUID="12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88" containerName="operator" containerID="cri-o://0c6b038c874feac00a0ea4a05fbd431f7014a4ed8fc71291c23ec972425e646d" gracePeriod=10 Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.145854 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-ddf2f"] Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.146315 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-ddf2f" podUID="e6d50b97-e5e2-426e-b881-dfb2077c0838" containerName="registry-server" containerID="cri-o://f6b5016e56cc05cedf6153ecd27ad5475f41f09c64ae9c2dcbdb602e86bd0445" gracePeriod=30 Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.170203 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s"] Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.173229 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zhd6s"] Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.195736 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74070831-862a-4d0a-83b0-4e3d64891601" path="/var/lib/kubelet/pods/74070831-862a-4d0a-83b0-4e3d64891601/volumes" Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.284257 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-xprqx" Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.374470 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2fkn\" (UniqueName: \"kubernetes.io/projected/12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88-kube-api-access-p2fkn\") pod \"12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88\" (UID: \"12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88\") " Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.392723 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88-kube-api-access-p2fkn" (OuterVolumeSpecName: "kube-api-access-p2fkn") pod "12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88" (UID: "12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88"). InnerVolumeSpecName "kube-api-access-p2fkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.476347 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2fkn\" (UniqueName: \"kubernetes.io/projected/12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88-kube-api-access-p2fkn\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.590426 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-ddf2f" Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.679115 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v42s4\" (UniqueName: \"kubernetes.io/projected/e6d50b97-e5e2-426e-b881-dfb2077c0838-kube-api-access-v42s4\") pod \"e6d50b97-e5e2-426e-b881-dfb2077c0838\" (UID: \"e6d50b97-e5e2-426e-b881-dfb2077c0838\") " Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.687106 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6d50b97-e5e2-426e-b881-dfb2077c0838-kube-api-access-v42s4" (OuterVolumeSpecName: "kube-api-access-v42s4") pod "e6d50b97-e5e2-426e-b881-dfb2077c0838" (UID: "e6d50b97-e5e2-426e-b881-dfb2077c0838"). InnerVolumeSpecName "kube-api-access-v42s4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.781222 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v42s4\" (UniqueName: \"kubernetes.io/projected/e6d50b97-e5e2-426e-b881-dfb2077c0838-kube-api-access-v42s4\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.823874 4714 generic.go:334] "Generic (PLEG): container finished" podID="e6d50b97-e5e2-426e-b881-dfb2077c0838" containerID="f6b5016e56cc05cedf6153ecd27ad5475f41f09c64ae9c2dcbdb602e86bd0445" exitCode=0 Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.823925 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-ddf2f" Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.823919 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-ddf2f" event={"ID":"e6d50b97-e5e2-426e-b881-dfb2077c0838","Type":"ContainerDied","Data":"f6b5016e56cc05cedf6153ecd27ad5475f41f09c64ae9c2dcbdb602e86bd0445"} Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.824052 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-ddf2f" event={"ID":"e6d50b97-e5e2-426e-b881-dfb2077c0838","Type":"ContainerDied","Data":"74d0d3468956b4f9354638adf80a846fa26535a7e74e370eeb721012c70bb9d7"} Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.824075 4714 scope.go:117] "RemoveContainer" containerID="f6b5016e56cc05cedf6153ecd27ad5475f41f09c64ae9c2dcbdb602e86bd0445" Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.826115 4714 generic.go:334] "Generic (PLEG): container finished" podID="12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88" containerID="0c6b038c874feac00a0ea4a05fbd431f7014a4ed8fc71291c23ec972425e646d" exitCode=0 Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.826144 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-xprqx" event={"ID":"12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88","Type":"ContainerDied","Data":"0c6b038c874feac00a0ea4a05fbd431f7014a4ed8fc71291c23ec972425e646d"} Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.826160 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-xprqx" event={"ID":"12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88","Type":"ContainerDied","Data":"9cef1a74877ad1a74c47eaad2e3c11a8681670f01de34b07c07c5549def07a12"} Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.827026 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-xprqx" Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.856765 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-ddf2f"] Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.860701 4714 scope.go:117] "RemoveContainer" containerID="f6b5016e56cc05cedf6153ecd27ad5475f41f09c64ae9c2dcbdb602e86bd0445" Jan 29 16:30:46 crc kubenswrapper[4714]: E0129 16:30:46.861096 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6b5016e56cc05cedf6153ecd27ad5475f41f09c64ae9c2dcbdb602e86bd0445\": container with ID starting with f6b5016e56cc05cedf6153ecd27ad5475f41f09c64ae9c2dcbdb602e86bd0445 not found: ID does not exist" containerID="f6b5016e56cc05cedf6153ecd27ad5475f41f09c64ae9c2dcbdb602e86bd0445" Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.861141 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6b5016e56cc05cedf6153ecd27ad5475f41f09c64ae9c2dcbdb602e86bd0445"} err="failed to get container status \"f6b5016e56cc05cedf6153ecd27ad5475f41f09c64ae9c2dcbdb602e86bd0445\": rpc error: code = NotFound desc = could not find container \"f6b5016e56cc05cedf6153ecd27ad5475f41f09c64ae9c2dcbdb602e86bd0445\": container with ID starting with f6b5016e56cc05cedf6153ecd27ad5475f41f09c64ae9c2dcbdb602e86bd0445 not found: ID does not exist" Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.861174 4714 scope.go:117] "RemoveContainer" containerID="0c6b038c874feac00a0ea4a05fbd431f7014a4ed8fc71291c23ec972425e646d" Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.863082 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-ddf2f"] Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.886037 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-xprqx"] Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.888030 4714 scope.go:117] "RemoveContainer" containerID="0c6b038c874feac00a0ea4a05fbd431f7014a4ed8fc71291c23ec972425e646d" Jan 29 16:30:46 crc kubenswrapper[4714]: E0129 16:30:46.888397 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c6b038c874feac00a0ea4a05fbd431f7014a4ed8fc71291c23ec972425e646d\": container with ID starting with 0c6b038c874feac00a0ea4a05fbd431f7014a4ed8fc71291c23ec972425e646d not found: ID does not exist" containerID="0c6b038c874feac00a0ea4a05fbd431f7014a4ed8fc71291c23ec972425e646d" Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.888430 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c6b038c874feac00a0ea4a05fbd431f7014a4ed8fc71291c23ec972425e646d"} err="failed to get container status \"0c6b038c874feac00a0ea4a05fbd431f7014a4ed8fc71291c23ec972425e646d\": rpc error: code = NotFound desc = could not find container \"0c6b038c874feac00a0ea4a05fbd431f7014a4ed8fc71291c23ec972425e646d\": container with ID starting with 0c6b038c874feac00a0ea4a05fbd431f7014a4ed8fc71291c23ec972425e646d not found: ID does not exist" Jan 29 16:30:46 crc kubenswrapper[4714]: I0129 16:30:46.889633 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-xprqx"] Jan 29 16:30:48 crc kubenswrapper[4714]: E0129 16:30:48.186689 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:30:48 crc kubenswrapper[4714]: I0129 16:30:48.194302 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88" path="/var/lib/kubelet/pods/12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88/volumes" Jan 29 16:30:48 crc kubenswrapper[4714]: I0129 16:30:48.195069 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6d50b97-e5e2-426e-b881-dfb2077c0838" path="/var/lib/kubelet/pods/e6d50b97-e5e2-426e-b881-dfb2077c0838/volumes" Jan 29 16:30:50 crc kubenswrapper[4714]: I0129 16:30:50.559631 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq"] Jan 29 16:30:50 crc kubenswrapper[4714]: I0129 16:30:50.560128 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" podUID="2779e724-225f-4a5f-9e2c-3b05fe08dff2" containerName="manager" containerID="cri-o://cec1cafa23793d2e5f4bd3af8e35a522a2c8cc5c802408ae2ad9896bd189471e" gracePeriod=10 Jan 29 16:30:50 crc kubenswrapper[4714]: I0129 16:30:50.850839 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-9xq82"] Jan 29 16:30:50 crc kubenswrapper[4714]: I0129 16:30:50.851460 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-9xq82" podUID="9dcd8561-aa17-46a8-b184-0495c320a33b" containerName="registry-server" containerID="cri-o://17fa7e5da52af363967b26a09672574689ebcabaf095f6c1443af9a04ea81254" gracePeriod=30 Jan 29 16:30:50 crc kubenswrapper[4714]: I0129 16:30:50.854811 4714 generic.go:334] "Generic (PLEG): container finished" podID="2779e724-225f-4a5f-9e2c-3b05fe08dff2" containerID="cec1cafa23793d2e5f4bd3af8e35a522a2c8cc5c802408ae2ad9896bd189471e" exitCode=0 Jan 29 16:30:50 crc kubenswrapper[4714]: I0129 16:30:50.854890 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" event={"ID":"2779e724-225f-4a5f-9e2c-3b05fe08dff2","Type":"ContainerDied","Data":"cec1cafa23793d2e5f4bd3af8e35a522a2c8cc5c802408ae2ad9896bd189471e"} Jan 29 16:30:50 crc kubenswrapper[4714]: I0129 16:30:50.921750 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs"] Jan 29 16:30:50 crc kubenswrapper[4714]: I0129 16:30:50.926703 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f757642kvs"] Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.154671 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.311498 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-9xq82" Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.341036 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2779e724-225f-4a5f-9e2c-3b05fe08dff2-apiservice-cert\") pod \"2779e724-225f-4a5f-9e2c-3b05fe08dff2\" (UID: \"2779e724-225f-4a5f-9e2c-3b05fe08dff2\") " Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.341115 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2779e724-225f-4a5f-9e2c-3b05fe08dff2-webhook-cert\") pod \"2779e724-225f-4a5f-9e2c-3b05fe08dff2\" (UID: \"2779e724-225f-4a5f-9e2c-3b05fe08dff2\") " Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.341170 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ts4kr\" (UniqueName: \"kubernetes.io/projected/9dcd8561-aa17-46a8-b184-0495c320a33b-kube-api-access-ts4kr\") pod \"9dcd8561-aa17-46a8-b184-0495c320a33b\" (UID: \"9dcd8561-aa17-46a8-b184-0495c320a33b\") " Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.341274 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jxvt\" (UniqueName: \"kubernetes.io/projected/2779e724-225f-4a5f-9e2c-3b05fe08dff2-kube-api-access-6jxvt\") pod \"2779e724-225f-4a5f-9e2c-3b05fe08dff2\" (UID: \"2779e724-225f-4a5f-9e2c-3b05fe08dff2\") " Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.345985 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dcd8561-aa17-46a8-b184-0495c320a33b-kube-api-access-ts4kr" (OuterVolumeSpecName: "kube-api-access-ts4kr") pod "9dcd8561-aa17-46a8-b184-0495c320a33b" (UID: "9dcd8561-aa17-46a8-b184-0495c320a33b"). InnerVolumeSpecName "kube-api-access-ts4kr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.346085 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2779e724-225f-4a5f-9e2c-3b05fe08dff2-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "2779e724-225f-4a5f-9e2c-3b05fe08dff2" (UID: "2779e724-225f-4a5f-9e2c-3b05fe08dff2"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.347195 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2779e724-225f-4a5f-9e2c-3b05fe08dff2-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "2779e724-225f-4a5f-9e2c-3b05fe08dff2" (UID: "2779e724-225f-4a5f-9e2c-3b05fe08dff2"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.354211 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2779e724-225f-4a5f-9e2c-3b05fe08dff2-kube-api-access-6jxvt" (OuterVolumeSpecName: "kube-api-access-6jxvt") pod "2779e724-225f-4a5f-9e2c-3b05fe08dff2" (UID: "2779e724-225f-4a5f-9e2c-3b05fe08dff2"). InnerVolumeSpecName "kube-api-access-6jxvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.442032 4714 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2779e724-225f-4a5f-9e2c-3b05fe08dff2-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.442073 4714 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2779e724-225f-4a5f-9e2c-3b05fe08dff2-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.442086 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ts4kr\" (UniqueName: \"kubernetes.io/projected/9dcd8561-aa17-46a8-b184-0495c320a33b-kube-api-access-ts4kr\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.442098 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jxvt\" (UniqueName: \"kubernetes.io/projected/2779e724-225f-4a5f-9e2c-3b05fe08dff2-kube-api-access-6jxvt\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.864636 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" event={"ID":"2779e724-225f-4a5f-9e2c-3b05fe08dff2","Type":"ContainerDied","Data":"a29d14b8a75b07c5bbb8bc497bdaef3d2bf58ebfc2cc63259d7bb78d82a74639"} Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.864710 4714 scope.go:117] "RemoveContainer" containerID="cec1cafa23793d2e5f4bd3af8e35a522a2c8cc5c802408ae2ad9896bd189471e" Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.864737 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq" Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.867055 4714 generic.go:334] "Generic (PLEG): container finished" podID="9dcd8561-aa17-46a8-b184-0495c320a33b" containerID="17fa7e5da52af363967b26a09672574689ebcabaf095f6c1443af9a04ea81254" exitCode=0 Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.867096 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-9xq82" event={"ID":"9dcd8561-aa17-46a8-b184-0495c320a33b","Type":"ContainerDied","Data":"17fa7e5da52af363967b26a09672574689ebcabaf095f6c1443af9a04ea81254"} Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.867125 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-9xq82" event={"ID":"9dcd8561-aa17-46a8-b184-0495c320a33b","Type":"ContainerDied","Data":"b25ca92cca2c102702aabf44744c607b11ac82433ac2ea8b9c60134c6952d1ea"} Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.867122 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-9xq82" Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.903761 4714 scope.go:117] "RemoveContainer" containerID="17fa7e5da52af363967b26a09672574689ebcabaf095f6c1443af9a04ea81254" Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.904644 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-9xq82"] Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.911645 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-9xq82"] Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.925250 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq"] Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.931756 4714 scope.go:117] "RemoveContainer" containerID="17fa7e5da52af363967b26a09672574689ebcabaf095f6c1443af9a04ea81254" Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.932198 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-controller-manager-66f4f5476c-xqnxq"] Jan 29 16:30:51 crc kubenswrapper[4714]: E0129 16:30:51.932296 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17fa7e5da52af363967b26a09672574689ebcabaf095f6c1443af9a04ea81254\": container with ID starting with 17fa7e5da52af363967b26a09672574689ebcabaf095f6c1443af9a04ea81254 not found: ID does not exist" containerID="17fa7e5da52af363967b26a09672574689ebcabaf095f6c1443af9a04ea81254" Jan 29 16:30:51 crc kubenswrapper[4714]: I0129 16:30:51.932333 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17fa7e5da52af363967b26a09672574689ebcabaf095f6c1443af9a04ea81254"} err="failed to get container status \"17fa7e5da52af363967b26a09672574689ebcabaf095f6c1443af9a04ea81254\": rpc error: code = NotFound desc = could not find container \"17fa7e5da52af363967b26a09672574689ebcabaf095f6c1443af9a04ea81254\": container with ID starting with 17fa7e5da52af363967b26a09672574689ebcabaf095f6c1443af9a04ea81254 not found: ID does not exist" Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.190920 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="096bd691-cca6-4566-b56c-7643e2feaef1" path="/var/lib/kubelet/pods/096bd691-cca6-4566-b56c-7643e2feaef1/volumes" Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.191565 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2779e724-225f-4a5f-9e2c-3b05fe08dff2" path="/var/lib/kubelet/pods/2779e724-225f-4a5f-9e2c-3b05fe08dff2/volumes" Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.192173 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dcd8561-aa17-46a8-b184-0495c320a33b" path="/var/lib/kubelet/pods/9dcd8561-aa17-46a8-b184-0495c320a33b/volumes" Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.446499 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn"] Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.446773 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" podUID="8771e447-1cf7-43f9-bfab-6c1afd7476dc" containerName="manager" containerID="cri-o://265cc03bd32fdd618d4ab75713d4df2f554d8e322232952f23e67ae7895f0208" gracePeriod=10 Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.743297 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-l6dkm"] Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.743487 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-l6dkm" podUID="d990dfb7-e078-4c7e-8e98-40b10f062a04" containerName="registry-server" containerID="cri-o://0020667ef371fcb5a3d00febffc3770f6cb20130544e17735a3ccff225db36b3" gracePeriod=30 Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.783883 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp"] Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.796654 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40zbdnp"] Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.885361 4714 generic.go:334] "Generic (PLEG): container finished" podID="d990dfb7-e078-4c7e-8e98-40b10f062a04" containerID="0020667ef371fcb5a3d00febffc3770f6cb20130544e17735a3ccff225db36b3" exitCode=0 Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.885480 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-l6dkm" event={"ID":"d990dfb7-e078-4c7e-8e98-40b10f062a04","Type":"ContainerDied","Data":"0020667ef371fcb5a3d00febffc3770f6cb20130544e17735a3ccff225db36b3"} Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.892864 4714 generic.go:334] "Generic (PLEG): container finished" podID="8771e447-1cf7-43f9-bfab-6c1afd7476dc" containerID="265cc03bd32fdd618d4ab75713d4df2f554d8e322232952f23e67ae7895f0208" exitCode=0 Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.892973 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" event={"ID":"8771e447-1cf7-43f9-bfab-6c1afd7476dc","Type":"ContainerDied","Data":"265cc03bd32fdd618d4ab75713d4df2f554d8e322232952f23e67ae7895f0208"} Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.893009 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" event={"ID":"8771e447-1cf7-43f9-bfab-6c1afd7476dc","Type":"ContainerDied","Data":"4ead9770f8d819bd3b2a9d514caf8dd18e92463bd9f126ba95d1cbb0e58fb71f"} Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.893057 4714 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ead9770f8d819bd3b2a9d514caf8dd18e92463bd9f126ba95d1cbb0e58fb71f" Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.903720 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.962992 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8771e447-1cf7-43f9-bfab-6c1afd7476dc-apiservice-cert\") pod \"8771e447-1cf7-43f9-bfab-6c1afd7476dc\" (UID: \"8771e447-1cf7-43f9-bfab-6c1afd7476dc\") " Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.963233 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mx2m7\" (UniqueName: \"kubernetes.io/projected/8771e447-1cf7-43f9-bfab-6c1afd7476dc-kube-api-access-mx2m7\") pod \"8771e447-1cf7-43f9-bfab-6c1afd7476dc\" (UID: \"8771e447-1cf7-43f9-bfab-6c1afd7476dc\") " Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.963279 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8771e447-1cf7-43f9-bfab-6c1afd7476dc-webhook-cert\") pod \"8771e447-1cf7-43f9-bfab-6c1afd7476dc\" (UID: \"8771e447-1cf7-43f9-bfab-6c1afd7476dc\") " Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.969370 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8771e447-1cf7-43f9-bfab-6c1afd7476dc-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "8771e447-1cf7-43f9-bfab-6c1afd7476dc" (UID: "8771e447-1cf7-43f9-bfab-6c1afd7476dc"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.969456 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8771e447-1cf7-43f9-bfab-6c1afd7476dc-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "8771e447-1cf7-43f9-bfab-6c1afd7476dc" (UID: "8771e447-1cf7-43f9-bfab-6c1afd7476dc"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:30:52 crc kubenswrapper[4714]: I0129 16:30:52.970823 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8771e447-1cf7-43f9-bfab-6c1afd7476dc-kube-api-access-mx2m7" (OuterVolumeSpecName: "kube-api-access-mx2m7") pod "8771e447-1cf7-43f9-bfab-6c1afd7476dc" (UID: "8771e447-1cf7-43f9-bfab-6c1afd7476dc"). InnerVolumeSpecName "kube-api-access-mx2m7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:53 crc kubenswrapper[4714]: I0129 16:30:53.064268 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mx2m7\" (UniqueName: \"kubernetes.io/projected/8771e447-1cf7-43f9-bfab-6c1afd7476dc-kube-api-access-mx2m7\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:53 crc kubenswrapper[4714]: I0129 16:30:53.064303 4714 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8771e447-1cf7-43f9-bfab-6c1afd7476dc-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:53 crc kubenswrapper[4714]: I0129 16:30:53.064314 4714 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8771e447-1cf7-43f9-bfab-6c1afd7476dc-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:53 crc kubenswrapper[4714]: I0129 16:30:53.103421 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-l6dkm" Jan 29 16:30:53 crc kubenswrapper[4714]: I0129 16:30:53.266791 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8r4sv\" (UniqueName: \"kubernetes.io/projected/d990dfb7-e078-4c7e-8e98-40b10f062a04-kube-api-access-8r4sv\") pod \"d990dfb7-e078-4c7e-8e98-40b10f062a04\" (UID: \"d990dfb7-e078-4c7e-8e98-40b10f062a04\") " Jan 29 16:30:53 crc kubenswrapper[4714]: I0129 16:30:53.269102 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d990dfb7-e078-4c7e-8e98-40b10f062a04-kube-api-access-8r4sv" (OuterVolumeSpecName: "kube-api-access-8r4sv") pod "d990dfb7-e078-4c7e-8e98-40b10f062a04" (UID: "d990dfb7-e078-4c7e-8e98-40b10f062a04"). InnerVolumeSpecName "kube-api-access-8r4sv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:30:53 crc kubenswrapper[4714]: I0129 16:30:53.368855 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8r4sv\" (UniqueName: \"kubernetes.io/projected/d990dfb7-e078-4c7e-8e98-40b10f062a04-kube-api-access-8r4sv\") on node \"crc\" DevicePath \"\"" Jan 29 16:30:53 crc kubenswrapper[4714]: I0129 16:30:53.900984 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-l6dkm" event={"ID":"d990dfb7-e078-4c7e-8e98-40b10f062a04","Type":"ContainerDied","Data":"cd19708e51ad0ae38749b06c635286e93f2554a3fecdeb70a36c4d4f40376c94"} Jan 29 16:30:53 crc kubenswrapper[4714]: I0129 16:30:53.901012 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-l6dkm" Jan 29 16:30:53 crc kubenswrapper[4714]: I0129 16:30:53.901043 4714 scope.go:117] "RemoveContainer" containerID="0020667ef371fcb5a3d00febffc3770f6cb20130544e17735a3ccff225db36b3" Jan 29 16:30:53 crc kubenswrapper[4714]: I0129 16:30:53.901005 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn" Jan 29 16:30:53 crc kubenswrapper[4714]: I0129 16:30:53.944036 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn"] Jan 29 16:30:53 crc kubenswrapper[4714]: I0129 16:30:53.950792 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7cc56799bb-ddchn"] Jan 29 16:30:53 crc kubenswrapper[4714]: I0129 16:30:53.959386 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-l6dkm"] Jan 29 16:30:53 crc kubenswrapper[4714]: I0129 16:30:53.962610 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-l6dkm"] Jan 29 16:30:54 crc kubenswrapper[4714]: I0129 16:30:54.194892 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8771e447-1cf7-43f9-bfab-6c1afd7476dc" path="/var/lib/kubelet/pods/8771e447-1cf7-43f9-bfab-6c1afd7476dc/volumes" Jan 29 16:30:54 crc kubenswrapper[4714]: I0129 16:30:54.196587 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="949d7185-7f54-44dd-9da9-3ed2c3c80e31" path="/var/lib/kubelet/pods/949d7185-7f54-44dd-9da9-3ed2c3c80e31/volumes" Jan 29 16:30:54 crc kubenswrapper[4714]: I0129 16:30:54.197322 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d990dfb7-e078-4c7e-8e98-40b10f062a04" path="/var/lib/kubelet/pods/d990dfb7-e078-4c7e-8e98-40b10f062a04/volumes" Jan 29 16:30:54 crc kubenswrapper[4714]: I0129 16:30:54.785094 4714 scope.go:117] "RemoveContainer" containerID="a2afc6d59d0b69e82adfd2f0ef885392ef80db9eed2c4295483123972e972c1a" Jan 29 16:30:54 crc kubenswrapper[4714]: I0129 16:30:54.822526 4714 scope.go:117] "RemoveContainer" containerID="bfc5719dfa5a492d30c2a6be943eb0655e0af9b3224bd4745d65e5929dc3407a" Jan 29 16:30:54 crc kubenswrapper[4714]: I0129 16:30:54.856101 4714 scope.go:117] "RemoveContainer" containerID="6ee4a0f3d055cfa18b2c55afd177524902b4dc64f544a61af9b1f46505e17336" Jan 29 16:30:54 crc kubenswrapper[4714]: I0129 16:30:54.879881 4714 scope.go:117] "RemoveContainer" containerID="265cc03bd32fdd618d4ab75713d4df2f554d8e322232952f23e67ae7895f0208" Jan 29 16:30:57 crc kubenswrapper[4714]: I0129 16:30:57.844276 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:30:57 crc kubenswrapper[4714]: I0129 16:30:57.846502 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:30:57 crc kubenswrapper[4714]: I0129 16:30:57.846711 4714 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:30:57 crc kubenswrapper[4714]: I0129 16:30:57.847809 4714 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6d286411f160a5fdbd13efa6bbfae544ec01e44f19ea6b8ff05d4ab9953a5f4a"} pod="openshift-machine-config-operator/machine-config-daemon-ppngk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 16:30:57 crc kubenswrapper[4714]: I0129 16:30:57.848595 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" containerID="cri-o://6d286411f160a5fdbd13efa6bbfae544ec01e44f19ea6b8ff05d4ab9953a5f4a" gracePeriod=600 Jan 29 16:30:58 crc kubenswrapper[4714]: I0129 16:30:58.946217 4714 generic.go:334] "Generic (PLEG): container finished" podID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerID="6d286411f160a5fdbd13efa6bbfae544ec01e44f19ea6b8ff05d4ab9953a5f4a" exitCode=0 Jan 29 16:30:58 crc kubenswrapper[4714]: I0129 16:30:58.946289 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" event={"ID":"c8c765f3-89eb-4077-8829-03e86eb0c90c","Type":"ContainerDied","Data":"6d286411f160a5fdbd13efa6bbfae544ec01e44f19ea6b8ff05d4ab9953a5f4a"} Jan 29 16:30:58 crc kubenswrapper[4714]: I0129 16:30:58.946393 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" event={"ID":"c8c765f3-89eb-4077-8829-03e86eb0c90c","Type":"ContainerStarted","Data":"28ae6797628a288c954e7899195453697f32c3fca947d19910c2ccc63b246a5b"} Jan 29 16:30:58 crc kubenswrapper[4714]: I0129 16:30:58.946474 4714 scope.go:117] "RemoveContainer" containerID="77045db0ac9dbee23fe648e58207222e15e50d5178fcc5cc7a606b4bbe2af7ec" Jan 29 16:31:02 crc kubenswrapper[4714]: E0129 16:31:02.186918 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802268 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qvwwf/must-gather-x95cf"] Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802673 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc06a535-6f60-438e-b52d-5dc90fae8c67" containerName="keystone-api" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802683 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc06a535-6f60-438e-b52d-5dc90fae8c67" containerName="keystone-api" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802698 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aa31790-7a3c-4a66-aace-c087c0221c6b" containerName="manager" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802704 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aa31790-7a3c-4a66-aace-c087c0221c6b" containerName="manager" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802717 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2779e724-225f-4a5f-9e2c-3b05fe08dff2" containerName="manager" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802722 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="2779e724-225f-4a5f-9e2c-3b05fe08dff2" containerName="manager" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802731 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8d336f2-b190-4e32-be3a-27fbf0e50a06" containerName="galera" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802738 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8d336f2-b190-4e32-be3a-27fbf0e50a06" containerName="galera" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802744 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea" containerName="memcached" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802750 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea" containerName="memcached" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802758 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8d336f2-b190-4e32-be3a-27fbf0e50a06" containerName="mysql-bootstrap" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802763 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8d336f2-b190-4e32-be3a-27fbf0e50a06" containerName="mysql-bootstrap" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802771 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55e23ac1-a89b-4689-a17d-bee875f7783e" containerName="rabbitmq" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802776 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="55e23ac1-a89b-4689-a17d-bee875f7783e" containerName="rabbitmq" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802785 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e367e739-45d9-4c71-82fa-ecda02da3277" containerName="galera" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802790 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="e367e739-45d9-4c71-82fa-ecda02da3277" containerName="galera" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802797 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e27f02c1-a7d5-4d49-838b-df5445720a07" containerName="galera" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802802 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="e27f02c1-a7d5-4d49-838b-df5445720a07" containerName="galera" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802810 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e27f02c1-a7d5-4d49-838b-df5445720a07" containerName="mysql-bootstrap" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802816 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="e27f02c1-a7d5-4d49-838b-df5445720a07" containerName="mysql-bootstrap" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802823 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d602ee5-4171-4dc7-9852-88c6019696e1" containerName="manager" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802829 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d602ee5-4171-4dc7-9852-88c6019696e1" containerName="manager" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802836 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87506df3-b56a-4598-8309-e865dc93cf53" containerName="registry-server" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802841 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="87506df3-b56a-4598-8309-e865dc93cf53" containerName="registry-server" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802848 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88" containerName="operator" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802853 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88" containerName="operator" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802860 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="112a1cde-5990-4140-97dd-c2bbc4f73197" containerName="mariadb-account-delete" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802868 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="112a1cde-5990-4140-97dd-c2bbc4f73197" containerName="mariadb-account-delete" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802877 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55e23ac1-a89b-4689-a17d-bee875f7783e" containerName="setup-container" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802882 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="55e23ac1-a89b-4689-a17d-bee875f7783e" containerName="setup-container" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802891 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="112a1cde-5990-4140-97dd-c2bbc4f73197" containerName="mariadb-account-delete" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802898 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="112a1cde-5990-4140-97dd-c2bbc4f73197" containerName="mariadb-account-delete" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802904 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d990dfb7-e078-4c7e-8e98-40b10f062a04" containerName="registry-server" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802910 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="d990dfb7-e078-4c7e-8e98-40b10f062a04" containerName="registry-server" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802918 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="668764e7-6295-4275-bcc9-24b680ec685f" containerName="registry-server" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802923 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="668764e7-6295-4275-bcc9-24b680ec685f" containerName="registry-server" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802944 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e367e739-45d9-4c71-82fa-ecda02da3277" containerName="mysql-bootstrap" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802949 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="e367e739-45d9-4c71-82fa-ecda02da3277" containerName="mysql-bootstrap" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802957 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6d50b97-e5e2-426e-b881-dfb2077c0838" containerName="registry-server" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802962 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6d50b97-e5e2-426e-b881-dfb2077c0838" containerName="registry-server" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802970 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dcd8561-aa17-46a8-b184-0495c320a33b" containerName="registry-server" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802975 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dcd8561-aa17-46a8-b184-0495c320a33b" containerName="registry-server" Jan 29 16:31:05 crc kubenswrapper[4714]: E0129 16:31:05.802984 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8771e447-1cf7-43f9-bfab-6c1afd7476dc" containerName="manager" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.802989 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="8771e447-1cf7-43f9-bfab-6c1afd7476dc" containerName="manager" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.803072 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="112a1cde-5990-4140-97dd-c2bbc4f73197" containerName="mariadb-account-delete" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.803082 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8d336f2-b190-4e32-be3a-27fbf0e50a06" containerName="galera" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.803091 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="12c4aba1-9cb7-4f12-bbd5-fee3ccfa0c88" containerName="operator" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.803100 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="e27f02c1-a7d5-4d49-838b-df5445720a07" containerName="galera" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.803107 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="55e23ac1-a89b-4689-a17d-bee875f7783e" containerName="rabbitmq" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.803114 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="e367e739-45d9-4c71-82fa-ecda02da3277" containerName="galera" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.803122 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0b0a6c4-5abe-4f28-9e81-56fcbf92f2ea" containerName="memcached" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.803129 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="668764e7-6295-4275-bcc9-24b680ec685f" containerName="registry-server" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.803136 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc06a535-6f60-438e-b52d-5dc90fae8c67" containerName="keystone-api" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.803146 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="87506df3-b56a-4598-8309-e865dc93cf53" containerName="registry-server" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.803151 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="8771e447-1cf7-43f9-bfab-6c1afd7476dc" containerName="manager" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.803159 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aa31790-7a3c-4a66-aace-c087c0221c6b" containerName="manager" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.803166 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6d50b97-e5e2-426e-b881-dfb2077c0838" containerName="registry-server" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.803174 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dcd8561-aa17-46a8-b184-0495c320a33b" containerName="registry-server" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.803180 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="2779e724-225f-4a5f-9e2c-3b05fe08dff2" containerName="manager" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.803186 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d602ee5-4171-4dc7-9852-88c6019696e1" containerName="manager" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.803193 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="d990dfb7-e078-4c7e-8e98-40b10f062a04" containerName="registry-server" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.803362 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="112a1cde-5990-4140-97dd-c2bbc4f73197" containerName="mariadb-account-delete" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.803708 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvwwf/must-gather-x95cf" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.807067 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-qvwwf"/"default-dockercfg-9z4tm" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.807099 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qvwwf"/"kube-root-ca.crt" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.807466 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qvwwf"/"openshift-service-ca.crt" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.810328 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qvwwf/must-gather-x95cf"] Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.949716 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm9z8\" (UniqueName: \"kubernetes.io/projected/7bff1b3a-4d70-4c22-ab8c-406d7e147f74-kube-api-access-lm9z8\") pod \"must-gather-x95cf\" (UID: \"7bff1b3a-4d70-4c22-ab8c-406d7e147f74\") " pod="openshift-must-gather-qvwwf/must-gather-x95cf" Jan 29 16:31:05 crc kubenswrapper[4714]: I0129 16:31:05.949867 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7bff1b3a-4d70-4c22-ab8c-406d7e147f74-must-gather-output\") pod \"must-gather-x95cf\" (UID: \"7bff1b3a-4d70-4c22-ab8c-406d7e147f74\") " pod="openshift-must-gather-qvwwf/must-gather-x95cf" Jan 29 16:31:06 crc kubenswrapper[4714]: I0129 16:31:06.051220 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7bff1b3a-4d70-4c22-ab8c-406d7e147f74-must-gather-output\") pod \"must-gather-x95cf\" (UID: \"7bff1b3a-4d70-4c22-ab8c-406d7e147f74\") " pod="openshift-must-gather-qvwwf/must-gather-x95cf" Jan 29 16:31:06 crc kubenswrapper[4714]: I0129 16:31:06.051264 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm9z8\" (UniqueName: \"kubernetes.io/projected/7bff1b3a-4d70-4c22-ab8c-406d7e147f74-kube-api-access-lm9z8\") pod \"must-gather-x95cf\" (UID: \"7bff1b3a-4d70-4c22-ab8c-406d7e147f74\") " pod="openshift-must-gather-qvwwf/must-gather-x95cf" Jan 29 16:31:06 crc kubenswrapper[4714]: I0129 16:31:06.051687 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7bff1b3a-4d70-4c22-ab8c-406d7e147f74-must-gather-output\") pod \"must-gather-x95cf\" (UID: \"7bff1b3a-4d70-4c22-ab8c-406d7e147f74\") " pod="openshift-must-gather-qvwwf/must-gather-x95cf" Jan 29 16:31:06 crc kubenswrapper[4714]: I0129 16:31:06.068767 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm9z8\" (UniqueName: \"kubernetes.io/projected/7bff1b3a-4d70-4c22-ab8c-406d7e147f74-kube-api-access-lm9z8\") pod \"must-gather-x95cf\" (UID: \"7bff1b3a-4d70-4c22-ab8c-406d7e147f74\") " pod="openshift-must-gather-qvwwf/must-gather-x95cf" Jan 29 16:31:06 crc kubenswrapper[4714]: I0129 16:31:06.118138 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvwwf/must-gather-x95cf" Jan 29 16:31:06 crc kubenswrapper[4714]: I0129 16:31:06.503193 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qvwwf/must-gather-x95cf"] Jan 29 16:31:07 crc kubenswrapper[4714]: I0129 16:31:07.007073 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qvwwf/must-gather-x95cf" event={"ID":"7bff1b3a-4d70-4c22-ab8c-406d7e147f74","Type":"ContainerStarted","Data":"507998147d3c1a984a1172e0e4c5307332cf30f147bf5ae992a56dd0e9660e72"} Jan 29 16:31:13 crc kubenswrapper[4714]: I0129 16:31:13.062293 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qvwwf/must-gather-x95cf" event={"ID":"7bff1b3a-4d70-4c22-ab8c-406d7e147f74","Type":"ContainerStarted","Data":"f6d6162656c64056efbbd383a714625735662747cd4e794e964ea668dbf260b4"} Jan 29 16:31:13 crc kubenswrapper[4714]: I0129 16:31:13.062899 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qvwwf/must-gather-x95cf" event={"ID":"7bff1b3a-4d70-4c22-ab8c-406d7e147f74","Type":"ContainerStarted","Data":"2547fd88c31e3900e045cc8b73057acb07ad9efeb1d1b5a29b5a911c036ec93d"} Jan 29 16:31:13 crc kubenswrapper[4714]: I0129 16:31:13.087622 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-qvwwf/must-gather-x95cf" podStartSLOduration=2.130600968 podStartE2EDuration="8.087603487s" podCreationTimestamp="2026-01-29 16:31:05 +0000 UTC" firstStartedPulling="2026-01-29 16:31:06.516701315 +0000 UTC m=+1273.037202485" lastFinishedPulling="2026-01-29 16:31:12.473703884 +0000 UTC m=+1278.994205004" observedRunningTime="2026-01-29 16:31:13.085108265 +0000 UTC m=+1279.605609385" watchObservedRunningTime="2026-01-29 16:31:13.087603487 +0000 UTC m=+1279.608104607" Jan 29 16:31:13 crc kubenswrapper[4714]: E0129 16:31:13.184884 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:31:25 crc kubenswrapper[4714]: E0129 16:31:25.186459 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:31:37 crc kubenswrapper[4714]: E0129 16:31:37.346802 4714 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 16:31:37 crc kubenswrapper[4714]: E0129 16:31:37.347581 4714 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tmm88,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-wdwq5_openshift-marketplace(8c12ad14-f878-42a1-a168-bad4026ec2dd): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 16:31:37 crc kubenswrapper[4714]: E0129 16:31:37.348775 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:31:50 crc kubenswrapper[4714]: E0129 16:31:50.185985 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:31:55 crc kubenswrapper[4714]: I0129 16:31:55.314801 4714 scope.go:117] "RemoveContainer" containerID="bf336482d3003324aec4b339b20443d78b6e477a15e7d92b43bd42b82e826811" Jan 29 16:31:55 crc kubenswrapper[4714]: I0129 16:31:55.353170 4714 scope.go:117] "RemoveContainer" containerID="1cd564e322fa134ca0b1e14f7b1c05c0de31a1e7c8f443cd6b64bbf340b9a6ae" Jan 29 16:31:55 crc kubenswrapper[4714]: I0129 16:31:55.377331 4714 scope.go:117] "RemoveContainer" containerID="8be88780f2ccaa67529f7c97f45a315a79167378ebf7e1fdcde84818ec246373" Jan 29 16:31:55 crc kubenswrapper[4714]: I0129 16:31:55.398850 4714 scope.go:117] "RemoveContainer" containerID="08be91d5ade94d67396b725df7d3290e5e0b4eed8f678b830b5a24bf0aefb822" Jan 29 16:31:55 crc kubenswrapper[4714]: I0129 16:31:55.424807 4714 scope.go:117] "RemoveContainer" containerID="3bc2a2089f6be296701e5d57c79cbb2a9a2dd560e4db7f4ea6460bad3386ed41" Jan 29 16:31:55 crc kubenswrapper[4714]: I0129 16:31:55.437198 4714 scope.go:117] "RemoveContainer" containerID="9a1454161993efe4d3d18d4c054f92025122ffa043acc6e49d820a2c93adec47" Jan 29 16:32:00 crc kubenswrapper[4714]: I0129 16:32:00.814956 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-sq9mx_8062d225-aa57-48df-bf28-2254ecc4f635/control-plane-machine-set-operator/0.log" Jan 29 16:32:00 crc kubenswrapper[4714]: I0129 16:32:00.963674 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-z4h55_bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92/kube-rbac-proxy/0.log" Jan 29 16:32:01 crc kubenswrapper[4714]: I0129 16:32:01.002172 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-z4h55_bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92/machine-api-operator/0.log" Jan 29 16:32:05 crc kubenswrapper[4714]: E0129 16:32:05.185653 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:32:19 crc kubenswrapper[4714]: E0129 16:32:19.185890 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:32:28 crc kubenswrapper[4714]: I0129 16:32:28.127634 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-m26zh_78b34628-144f-416a-b493-15ba445caa48/controller/0.log" Jan 29 16:32:28 crc kubenswrapper[4714]: I0129 16:32:28.133581 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-m26zh_78b34628-144f-416a-b493-15ba445caa48/kube-rbac-proxy/0.log" Jan 29 16:32:28 crc kubenswrapper[4714]: I0129 16:32:28.374408 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-frr-files/0.log" Jan 29 16:32:28 crc kubenswrapper[4714]: I0129 16:32:28.487912 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-frr-files/0.log" Jan 29 16:32:28 crc kubenswrapper[4714]: I0129 16:32:28.519802 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-reloader/0.log" Jan 29 16:32:28 crc kubenswrapper[4714]: I0129 16:32:28.569503 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-metrics/0.log" Jan 29 16:32:28 crc kubenswrapper[4714]: I0129 16:32:28.585684 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-reloader/0.log" Jan 29 16:32:28 crc kubenswrapper[4714]: I0129 16:32:28.740717 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-reloader/0.log" Jan 29 16:32:28 crc kubenswrapper[4714]: I0129 16:32:28.747540 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-frr-files/0.log" Jan 29 16:32:28 crc kubenswrapper[4714]: I0129 16:32:28.782249 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-metrics/0.log" Jan 29 16:32:28 crc kubenswrapper[4714]: I0129 16:32:28.784365 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-metrics/0.log" Jan 29 16:32:28 crc kubenswrapper[4714]: I0129 16:32:28.950366 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-metrics/0.log" Jan 29 16:32:28 crc kubenswrapper[4714]: I0129 16:32:28.952061 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-reloader/0.log" Jan 29 16:32:28 crc kubenswrapper[4714]: I0129 16:32:28.963845 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-frr-files/0.log" Jan 29 16:32:28 crc kubenswrapper[4714]: I0129 16:32:28.991200 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/controller/0.log" Jan 29 16:32:29 crc kubenswrapper[4714]: I0129 16:32:29.102396 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/frr-metrics/0.log" Jan 29 16:32:29 crc kubenswrapper[4714]: I0129 16:32:29.149036 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/kube-rbac-proxy/0.log" Jan 29 16:32:29 crc kubenswrapper[4714]: I0129 16:32:29.184042 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/kube-rbac-proxy-frr/0.log" Jan 29 16:32:29 crc kubenswrapper[4714]: I0129 16:32:29.322733 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/reloader/0.log" Jan 29 16:32:29 crc kubenswrapper[4714]: I0129 16:32:29.391069 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-kk79r_9bbfcf92-8a27-4ba0-9017-7c36906791c8/frr-k8s-webhook-server/0.log" Jan 29 16:32:29 crc kubenswrapper[4714]: I0129 16:32:29.513901 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/frr/0.log" Jan 29 16:32:29 crc kubenswrapper[4714]: I0129 16:32:29.588502 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-586b87b897-zpr4q_432a4f98-877c-4f7a-b2b0-ce273a77450a/manager/0.log" Jan 29 16:32:29 crc kubenswrapper[4714]: I0129 16:32:29.834968 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7df7c8d444-xs67n_ffe179b8-a1c8-430b-94f5-920aacf0defe/webhook-server/0.log" Jan 29 16:32:29 crc kubenswrapper[4714]: I0129 16:32:29.874589 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7mmsh_813f735d-8336-49e9-b018-e6dbf74ddc99/kube-rbac-proxy/0.log" Jan 29 16:32:29 crc kubenswrapper[4714]: I0129 16:32:29.989308 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7mmsh_813f735d-8336-49e9-b018-e6dbf74ddc99/speaker/0.log" Jan 29 16:32:31 crc kubenswrapper[4714]: E0129 16:32:31.186229 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:32:43 crc kubenswrapper[4714]: E0129 16:32:43.187799 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:32:55 crc kubenswrapper[4714]: I0129 16:32:55.485290 4714 scope.go:117] "RemoveContainer" containerID="e9b290e5fae1b9ebd91874f0c7f54baf70c50604b0924e9b333424187e1578aa" Jan 29 16:32:55 crc kubenswrapper[4714]: I0129 16:32:55.509809 4714 scope.go:117] "RemoveContainer" containerID="ae1e8fd69fe054dba679bc3d816ff2486311c7fd767f49bb3c77b8a2f9da9054" Jan 29 16:32:55 crc kubenswrapper[4714]: I0129 16:32:55.526246 4714 scope.go:117] "RemoveContainer" containerID="cdd9e5087dbe100b4200dc045cc6536b7bd5644b604c1ad48cd724f12116a2d5" Jan 29 16:32:55 crc kubenswrapper[4714]: I0129 16:32:55.575304 4714 scope.go:117] "RemoveContainer" containerID="a293414bfe6fecbd34f2097ce525abe6d50f764495aa3e4545c1f2cdb4d889ff" Jan 29 16:32:55 crc kubenswrapper[4714]: I0129 16:32:55.947143 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_c184c6f2-1af5-4f70-9251-6beb2baae06b/util/0.log" Jan 29 16:32:56 crc kubenswrapper[4714]: I0129 16:32:56.115962 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_c184c6f2-1af5-4f70-9251-6beb2baae06b/util/0.log" Jan 29 16:32:56 crc kubenswrapper[4714]: I0129 16:32:56.135360 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_c184c6f2-1af5-4f70-9251-6beb2baae06b/pull/0.log" Jan 29 16:32:56 crc kubenswrapper[4714]: I0129 16:32:56.139517 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_c184c6f2-1af5-4f70-9251-6beb2baae06b/pull/0.log" Jan 29 16:32:56 crc kubenswrapper[4714]: I0129 16:32:56.331893 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_c184c6f2-1af5-4f70-9251-6beb2baae06b/extract/0.log" Jan 29 16:32:56 crc kubenswrapper[4714]: I0129 16:32:56.351773 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_c184c6f2-1af5-4f70-9251-6beb2baae06b/util/0.log" Jan 29 16:32:56 crc kubenswrapper[4714]: I0129 16:32:56.360351 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_c184c6f2-1af5-4f70-9251-6beb2baae06b/pull/0.log" Jan 29 16:32:56 crc kubenswrapper[4714]: I0129 16:32:56.526353 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slsxz_16cb244c-6c63-47e6-a312-ba33ab4d4899/extract-utilities/0.log" Jan 29 16:32:56 crc kubenswrapper[4714]: I0129 16:32:56.663542 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slsxz_16cb244c-6c63-47e6-a312-ba33ab4d4899/extract-content/0.log" Jan 29 16:32:56 crc kubenswrapper[4714]: I0129 16:32:56.714268 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slsxz_16cb244c-6c63-47e6-a312-ba33ab4d4899/extract-utilities/0.log" Jan 29 16:32:56 crc kubenswrapper[4714]: I0129 16:32:56.726146 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slsxz_16cb244c-6c63-47e6-a312-ba33ab4d4899/extract-content/0.log" Jan 29 16:32:56 crc kubenswrapper[4714]: I0129 16:32:56.881020 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slsxz_16cb244c-6c63-47e6-a312-ba33ab4d4899/extract-utilities/0.log" Jan 29 16:32:56 crc kubenswrapper[4714]: I0129 16:32:56.915048 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slsxz_16cb244c-6c63-47e6-a312-ba33ab4d4899/extract-content/0.log" Jan 29 16:32:57 crc kubenswrapper[4714]: I0129 16:32:57.224467 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slsxz_16cb244c-6c63-47e6-a312-ba33ab4d4899/registry-server/0.log" Jan 29 16:32:57 crc kubenswrapper[4714]: I0129 16:32:57.322891 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ndx6p_ca655e22-8f97-4e9e-b115-734ae1af7d50/extract-utilities/0.log" Jan 29 16:32:57 crc kubenswrapper[4714]: I0129 16:32:57.481342 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ndx6p_ca655e22-8f97-4e9e-b115-734ae1af7d50/extract-content/0.log" Jan 29 16:32:57 crc kubenswrapper[4714]: I0129 16:32:57.504306 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ndx6p_ca655e22-8f97-4e9e-b115-734ae1af7d50/extract-content/0.log" Jan 29 16:32:57 crc kubenswrapper[4714]: I0129 16:32:57.524618 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ndx6p_ca655e22-8f97-4e9e-b115-734ae1af7d50/extract-utilities/0.log" Jan 29 16:32:57 crc kubenswrapper[4714]: I0129 16:32:57.665873 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ndx6p_ca655e22-8f97-4e9e-b115-734ae1af7d50/extract-utilities/0.log" Jan 29 16:32:57 crc kubenswrapper[4714]: I0129 16:32:57.690782 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ndx6p_ca655e22-8f97-4e9e-b115-734ae1af7d50/extract-content/0.log" Jan 29 16:32:57 crc kubenswrapper[4714]: I0129 16:32:57.888133 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-7rvrl_2696757f-83ca-42df-9855-f76adeee02bb/marketplace-operator/0.log" Jan 29 16:32:58 crc kubenswrapper[4714]: I0129 16:32:58.005365 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6gkpz_04dba3a0-a89b-48c5-97ef-e5660d1ae7bb/extract-utilities/0.log" Jan 29 16:32:58 crc kubenswrapper[4714]: I0129 16:32:58.013305 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ndx6p_ca655e22-8f97-4e9e-b115-734ae1af7d50/registry-server/0.log" Jan 29 16:32:58 crc kubenswrapper[4714]: I0129 16:32:58.106072 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6gkpz_04dba3a0-a89b-48c5-97ef-e5660d1ae7bb/extract-content/0.log" Jan 29 16:32:58 crc kubenswrapper[4714]: I0129 16:32:58.139593 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6gkpz_04dba3a0-a89b-48c5-97ef-e5660d1ae7bb/extract-utilities/0.log" Jan 29 16:32:58 crc kubenswrapper[4714]: I0129 16:32:58.168486 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6gkpz_04dba3a0-a89b-48c5-97ef-e5660d1ae7bb/extract-content/0.log" Jan 29 16:32:58 crc kubenswrapper[4714]: E0129 16:32:58.185509 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:32:58 crc kubenswrapper[4714]: I0129 16:32:58.316617 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6gkpz_04dba3a0-a89b-48c5-97ef-e5660d1ae7bb/extract-utilities/0.log" Jan 29 16:32:58 crc kubenswrapper[4714]: I0129 16:32:58.383124 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6gkpz_04dba3a0-a89b-48c5-97ef-e5660d1ae7bb/extract-content/0.log" Jan 29 16:32:58 crc kubenswrapper[4714]: I0129 16:32:58.436783 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6gkpz_04dba3a0-a89b-48c5-97ef-e5660d1ae7bb/registry-server/0.log" Jan 29 16:32:58 crc kubenswrapper[4714]: I0129 16:32:58.484078 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wdwq5_8c12ad14-f878-42a1-a168-bad4026ec2dd/extract-utilities/0.log" Jan 29 16:32:58 crc kubenswrapper[4714]: I0129 16:32:58.624134 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wdwq5_8c12ad14-f878-42a1-a168-bad4026ec2dd/extract-utilities/0.log" Jan 29 16:32:58 crc kubenswrapper[4714]: I0129 16:32:58.860351 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wdwq5_8c12ad14-f878-42a1-a168-bad4026ec2dd/extract-utilities/0.log" Jan 29 16:32:58 crc kubenswrapper[4714]: I0129 16:32:58.981109 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-knxc8_de6c9fbd-8657-4434-bff5-468276791466/extract-utilities/0.log" Jan 29 16:32:59 crc kubenswrapper[4714]: I0129 16:32:59.209178 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-knxc8_de6c9fbd-8657-4434-bff5-468276791466/extract-content/0.log" Jan 29 16:32:59 crc kubenswrapper[4714]: I0129 16:32:59.273106 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-knxc8_de6c9fbd-8657-4434-bff5-468276791466/extract-utilities/0.log" Jan 29 16:32:59 crc kubenswrapper[4714]: I0129 16:32:59.274153 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-knxc8_de6c9fbd-8657-4434-bff5-468276791466/extract-content/0.log" Jan 29 16:32:59 crc kubenswrapper[4714]: I0129 16:32:59.489573 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-knxc8_de6c9fbd-8657-4434-bff5-468276791466/extract-utilities/0.log" Jan 29 16:32:59 crc kubenswrapper[4714]: I0129 16:32:59.507868 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-knxc8_de6c9fbd-8657-4434-bff5-468276791466/extract-content/0.log" Jan 29 16:32:59 crc kubenswrapper[4714]: I0129 16:32:59.666473 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-knxc8_de6c9fbd-8657-4434-bff5-468276791466/registry-server/0.log" Jan 29 16:33:13 crc kubenswrapper[4714]: E0129 16:33:13.187387 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:33:27 crc kubenswrapper[4714]: E0129 16:33:27.186528 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:33:27 crc kubenswrapper[4714]: I0129 16:33:27.844241 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:33:27 crc kubenswrapper[4714]: I0129 16:33:27.844322 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:33:41 crc kubenswrapper[4714]: E0129 16:33:41.187982 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:33:52 crc kubenswrapper[4714]: E0129 16:33:52.186843 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:33:55 crc kubenswrapper[4714]: I0129 16:33:55.648525 4714 scope.go:117] "RemoveContainer" containerID="a675f4f90dd437578a32683cafb8b1908c7b80f63189aec46569a29c2add56c0" Jan 29 16:33:55 crc kubenswrapper[4714]: I0129 16:33:55.680598 4714 scope.go:117] "RemoveContainer" containerID="02c31eb5896b8dc80e78bcd830c0e2f150e33491e2a451310581ca2dc793d036" Jan 29 16:33:55 crc kubenswrapper[4714]: I0129 16:33:55.719380 4714 scope.go:117] "RemoveContainer" containerID="57df90298e52fbd874f183f4f349569b890f49a89aba1583825eacf13909d613" Jan 29 16:33:55 crc kubenswrapper[4714]: I0129 16:33:55.758613 4714 scope.go:117] "RemoveContainer" containerID="6cf56f6dac5db6cefc7926b5a24bd8c2963224c5d6e15dd78662ec20f0cf0141" Jan 29 16:33:55 crc kubenswrapper[4714]: I0129 16:33:55.828027 4714 scope.go:117] "RemoveContainer" containerID="6f5e7b1c14376dcb0e62786390737b48cb04d2a05079d8160df0f9dedc56dfc8" Jan 29 16:33:55 crc kubenswrapper[4714]: I0129 16:33:55.846612 4714 scope.go:117] "RemoveContainer" containerID="8f755ca88cec23079c8fcc603a70054716d3829ccb1c6da9ff0f5feff88b5796" Jan 29 16:33:55 crc kubenswrapper[4714]: I0129 16:33:55.873825 4714 scope.go:117] "RemoveContainer" containerID="958fc7292af56ecfa7d5c5a7066233d1295c2b0c82a6e8e4646901914aabf005" Jan 29 16:33:57 crc kubenswrapper[4714]: I0129 16:33:57.844857 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:33:57 crc kubenswrapper[4714]: I0129 16:33:57.844987 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:34:06 crc kubenswrapper[4714]: E0129 16:34:06.191537 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:34:11 crc kubenswrapper[4714]: I0129 16:34:11.321241 4714 generic.go:334] "Generic (PLEG): container finished" podID="7bff1b3a-4d70-4c22-ab8c-406d7e147f74" containerID="2547fd88c31e3900e045cc8b73057acb07ad9efeb1d1b5a29b5a911c036ec93d" exitCode=0 Jan 29 16:34:11 crc kubenswrapper[4714]: I0129 16:34:11.321361 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qvwwf/must-gather-x95cf" event={"ID":"7bff1b3a-4d70-4c22-ab8c-406d7e147f74","Type":"ContainerDied","Data":"2547fd88c31e3900e045cc8b73057acb07ad9efeb1d1b5a29b5a911c036ec93d"} Jan 29 16:34:11 crc kubenswrapper[4714]: I0129 16:34:11.322191 4714 scope.go:117] "RemoveContainer" containerID="2547fd88c31e3900e045cc8b73057acb07ad9efeb1d1b5a29b5a911c036ec93d" Jan 29 16:34:11 crc kubenswrapper[4714]: I0129 16:34:11.435579 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qvwwf_must-gather-x95cf_7bff1b3a-4d70-4c22-ab8c-406d7e147f74/gather/0.log" Jan 29 16:34:18 crc kubenswrapper[4714]: I0129 16:34:18.171613 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qvwwf/must-gather-x95cf"] Jan 29 16:34:18 crc kubenswrapper[4714]: I0129 16:34:18.172293 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-qvwwf/must-gather-x95cf" podUID="7bff1b3a-4d70-4c22-ab8c-406d7e147f74" containerName="copy" containerID="cri-o://f6d6162656c64056efbbd383a714625735662747cd4e794e964ea668dbf260b4" gracePeriod=2 Jan 29 16:34:18 crc kubenswrapper[4714]: I0129 16:34:18.180887 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qvwwf/must-gather-x95cf"] Jan 29 16:34:18 crc kubenswrapper[4714]: I0129 16:34:18.396228 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qvwwf_must-gather-x95cf_7bff1b3a-4d70-4c22-ab8c-406d7e147f74/copy/0.log" Jan 29 16:34:18 crc kubenswrapper[4714]: I0129 16:34:18.396579 4714 generic.go:334] "Generic (PLEG): container finished" podID="7bff1b3a-4d70-4c22-ab8c-406d7e147f74" containerID="f6d6162656c64056efbbd383a714625735662747cd4e794e964ea668dbf260b4" exitCode=143 Jan 29 16:34:18 crc kubenswrapper[4714]: I0129 16:34:18.538330 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qvwwf_must-gather-x95cf_7bff1b3a-4d70-4c22-ab8c-406d7e147f74/copy/0.log" Jan 29 16:34:18 crc kubenswrapper[4714]: I0129 16:34:18.543159 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvwwf/must-gather-x95cf" Jan 29 16:34:18 crc kubenswrapper[4714]: I0129 16:34:18.595557 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm9z8\" (UniqueName: \"kubernetes.io/projected/7bff1b3a-4d70-4c22-ab8c-406d7e147f74-kube-api-access-lm9z8\") pod \"7bff1b3a-4d70-4c22-ab8c-406d7e147f74\" (UID: \"7bff1b3a-4d70-4c22-ab8c-406d7e147f74\") " Jan 29 16:34:18 crc kubenswrapper[4714]: I0129 16:34:18.595708 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7bff1b3a-4d70-4c22-ab8c-406d7e147f74-must-gather-output\") pod \"7bff1b3a-4d70-4c22-ab8c-406d7e147f74\" (UID: \"7bff1b3a-4d70-4c22-ab8c-406d7e147f74\") " Jan 29 16:34:18 crc kubenswrapper[4714]: I0129 16:34:18.605702 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bff1b3a-4d70-4c22-ab8c-406d7e147f74-kube-api-access-lm9z8" (OuterVolumeSpecName: "kube-api-access-lm9z8") pod "7bff1b3a-4d70-4c22-ab8c-406d7e147f74" (UID: "7bff1b3a-4d70-4c22-ab8c-406d7e147f74"). InnerVolumeSpecName "kube-api-access-lm9z8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:34:18 crc kubenswrapper[4714]: I0129 16:34:18.653318 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bff1b3a-4d70-4c22-ab8c-406d7e147f74-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "7bff1b3a-4d70-4c22-ab8c-406d7e147f74" (UID: "7bff1b3a-4d70-4c22-ab8c-406d7e147f74"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:34:18 crc kubenswrapper[4714]: I0129 16:34:18.697459 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lm9z8\" (UniqueName: \"kubernetes.io/projected/7bff1b3a-4d70-4c22-ab8c-406d7e147f74-kube-api-access-lm9z8\") on node \"crc\" DevicePath \"\"" Jan 29 16:34:18 crc kubenswrapper[4714]: I0129 16:34:18.697552 4714 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7bff1b3a-4d70-4c22-ab8c-406d7e147f74-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 29 16:34:19 crc kubenswrapper[4714]: I0129 16:34:19.407354 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qvwwf_must-gather-x95cf_7bff1b3a-4d70-4c22-ab8c-406d7e147f74/copy/0.log" Jan 29 16:34:19 crc kubenswrapper[4714]: I0129 16:34:19.408544 4714 scope.go:117] "RemoveContainer" containerID="f6d6162656c64056efbbd383a714625735662747cd4e794e964ea668dbf260b4" Jan 29 16:34:19 crc kubenswrapper[4714]: I0129 16:34:19.408724 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvwwf/must-gather-x95cf" Jan 29 16:34:19 crc kubenswrapper[4714]: I0129 16:34:19.435396 4714 scope.go:117] "RemoveContainer" containerID="2547fd88c31e3900e045cc8b73057acb07ad9efeb1d1b5a29b5a911c036ec93d" Jan 29 16:34:20 crc kubenswrapper[4714]: I0129 16:34:20.189835 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bff1b3a-4d70-4c22-ab8c-406d7e147f74" path="/var/lib/kubelet/pods/7bff1b3a-4d70-4c22-ab8c-406d7e147f74/volumes" Jan 29 16:34:21 crc kubenswrapper[4714]: E0129 16:34:21.186683 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:34:27 crc kubenswrapper[4714]: I0129 16:34:27.844851 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:34:27 crc kubenswrapper[4714]: I0129 16:34:27.845280 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:34:27 crc kubenswrapper[4714]: I0129 16:34:27.845340 4714 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:34:27 crc kubenswrapper[4714]: I0129 16:34:27.846170 4714 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"28ae6797628a288c954e7899195453697f32c3fca947d19910c2ccc63b246a5b"} pod="openshift-machine-config-operator/machine-config-daemon-ppngk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 16:34:27 crc kubenswrapper[4714]: I0129 16:34:27.846267 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" containerID="cri-o://28ae6797628a288c954e7899195453697f32c3fca947d19910c2ccc63b246a5b" gracePeriod=600 Jan 29 16:34:28 crc kubenswrapper[4714]: I0129 16:34:28.472035 4714 generic.go:334] "Generic (PLEG): container finished" podID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerID="28ae6797628a288c954e7899195453697f32c3fca947d19910c2ccc63b246a5b" exitCode=0 Jan 29 16:34:28 crc kubenswrapper[4714]: I0129 16:34:28.472091 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" event={"ID":"c8c765f3-89eb-4077-8829-03e86eb0c90c","Type":"ContainerDied","Data":"28ae6797628a288c954e7899195453697f32c3fca947d19910c2ccc63b246a5b"} Jan 29 16:34:28 crc kubenswrapper[4714]: I0129 16:34:28.472957 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" event={"ID":"c8c765f3-89eb-4077-8829-03e86eb0c90c","Type":"ContainerStarted","Data":"0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1"} Jan 29 16:34:28 crc kubenswrapper[4714]: I0129 16:34:28.473009 4714 scope.go:117] "RemoveContainer" containerID="6d286411f160a5fdbd13efa6bbfae544ec01e44f19ea6b8ff05d4ab9953a5f4a" Jan 29 16:34:33 crc kubenswrapper[4714]: E0129 16:34:33.187303 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:34:46 crc kubenswrapper[4714]: E0129 16:34:46.188270 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:34:55 crc kubenswrapper[4714]: I0129 16:34:55.957727 4714 scope.go:117] "RemoveContainer" containerID="cd8982aadd49edb0050578e4754c053be8b4e593ef390dafa6884ec2cec1fd5d" Jan 29 16:34:55 crc kubenswrapper[4714]: I0129 16:34:55.987271 4714 scope.go:117] "RemoveContainer" containerID="bc9178e686ab88b7e47825dd5faad2c6f1b972c479a40bdd9847878e376e9b8c" Jan 29 16:34:56 crc kubenswrapper[4714]: I0129 16:34:56.025344 4714 scope.go:117] "RemoveContainer" containerID="83eeaf58ca15604fd125219fc4be09b86cbb0308fb89b0438c4ada31625917b5" Jan 29 16:34:58 crc kubenswrapper[4714]: E0129 16:34:58.185861 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:35:09 crc kubenswrapper[4714]: E0129 16:35:09.188200 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:35:24 crc kubenswrapper[4714]: E0129 16:35:24.189908 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:35:37 crc kubenswrapper[4714]: E0129 16:35:37.187318 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:35:49 crc kubenswrapper[4714]: E0129 16:35:49.188756 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:35:56 crc kubenswrapper[4714]: I0129 16:35:56.125002 4714 scope.go:117] "RemoveContainer" containerID="06b2f42073c400aedcdcd2bb5ec2c469776d6fe3bfbe8ce136fb5ad93196a673" Jan 29 16:35:56 crc kubenswrapper[4714]: I0129 16:35:56.160711 4714 scope.go:117] "RemoveContainer" containerID="bbd0c612c943b6ec94ed1610e6f355545ecaf388e165bbb67cb2c301b434e969" Jan 29 16:36:02 crc kubenswrapper[4714]: E0129 16:36:02.186675 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:36:14 crc kubenswrapper[4714]: E0129 16:36:14.190396 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:36:28 crc kubenswrapper[4714]: E0129 16:36:28.187350 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" Jan 29 16:36:41 crc kubenswrapper[4714]: I0129 16:36:41.551579 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hqc4j/must-gather-684p8"] Jan 29 16:36:41 crc kubenswrapper[4714]: E0129 16:36:41.552865 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bff1b3a-4d70-4c22-ab8c-406d7e147f74" containerName="gather" Jan 29 16:36:41 crc kubenswrapper[4714]: I0129 16:36:41.553121 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bff1b3a-4d70-4c22-ab8c-406d7e147f74" containerName="gather" Jan 29 16:36:41 crc kubenswrapper[4714]: E0129 16:36:41.553159 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bff1b3a-4d70-4c22-ab8c-406d7e147f74" containerName="copy" Jan 29 16:36:41 crc kubenswrapper[4714]: I0129 16:36:41.553177 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bff1b3a-4d70-4c22-ab8c-406d7e147f74" containerName="copy" Jan 29 16:36:41 crc kubenswrapper[4714]: I0129 16:36:41.553429 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bff1b3a-4d70-4c22-ab8c-406d7e147f74" containerName="copy" Jan 29 16:36:41 crc kubenswrapper[4714]: I0129 16:36:41.553470 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bff1b3a-4d70-4c22-ab8c-406d7e147f74" containerName="gather" Jan 29 16:36:41 crc kubenswrapper[4714]: I0129 16:36:41.554775 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hqc4j/must-gather-684p8" Jan 29 16:36:41 crc kubenswrapper[4714]: I0129 16:36:41.557511 4714 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-hqc4j"/"default-dockercfg-jhg4d" Jan 29 16:36:41 crc kubenswrapper[4714]: I0129 16:36:41.558119 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-hqc4j"/"kube-root-ca.crt" Jan 29 16:36:41 crc kubenswrapper[4714]: I0129 16:36:41.558158 4714 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-hqc4j"/"openshift-service-ca.crt" Jan 29 16:36:41 crc kubenswrapper[4714]: I0129 16:36:41.561835 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-hqc4j/must-gather-684p8"] Jan 29 16:36:41 crc kubenswrapper[4714]: I0129 16:36:41.708355 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jf26\" (UniqueName: \"kubernetes.io/projected/ff3dd7de-0ec3-4550-ba38-0d67405a2671-kube-api-access-8jf26\") pod \"must-gather-684p8\" (UID: \"ff3dd7de-0ec3-4550-ba38-0d67405a2671\") " pod="openshift-must-gather-hqc4j/must-gather-684p8" Jan 29 16:36:41 crc kubenswrapper[4714]: I0129 16:36:41.708539 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ff3dd7de-0ec3-4550-ba38-0d67405a2671-must-gather-output\") pod \"must-gather-684p8\" (UID: \"ff3dd7de-0ec3-4550-ba38-0d67405a2671\") " pod="openshift-must-gather-hqc4j/must-gather-684p8" Jan 29 16:36:41 crc kubenswrapper[4714]: I0129 16:36:41.809771 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jf26\" (UniqueName: \"kubernetes.io/projected/ff3dd7de-0ec3-4550-ba38-0d67405a2671-kube-api-access-8jf26\") pod \"must-gather-684p8\" (UID: \"ff3dd7de-0ec3-4550-ba38-0d67405a2671\") " pod="openshift-must-gather-hqc4j/must-gather-684p8" Jan 29 16:36:41 crc kubenswrapper[4714]: I0129 16:36:41.809863 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ff3dd7de-0ec3-4550-ba38-0d67405a2671-must-gather-output\") pod \"must-gather-684p8\" (UID: \"ff3dd7de-0ec3-4550-ba38-0d67405a2671\") " pod="openshift-must-gather-hqc4j/must-gather-684p8" Jan 29 16:36:41 crc kubenswrapper[4714]: I0129 16:36:41.810556 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ff3dd7de-0ec3-4550-ba38-0d67405a2671-must-gather-output\") pod \"must-gather-684p8\" (UID: \"ff3dd7de-0ec3-4550-ba38-0d67405a2671\") " pod="openshift-must-gather-hqc4j/must-gather-684p8" Jan 29 16:36:41 crc kubenswrapper[4714]: I0129 16:36:41.828323 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jf26\" (UniqueName: \"kubernetes.io/projected/ff3dd7de-0ec3-4550-ba38-0d67405a2671-kube-api-access-8jf26\") pod \"must-gather-684p8\" (UID: \"ff3dd7de-0ec3-4550-ba38-0d67405a2671\") " pod="openshift-must-gather-hqc4j/must-gather-684p8" Jan 29 16:36:41 crc kubenswrapper[4714]: I0129 16:36:41.876395 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hqc4j/must-gather-684p8" Jan 29 16:36:42 crc kubenswrapper[4714]: I0129 16:36:42.137904 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-hqc4j/must-gather-684p8"] Jan 29 16:36:42 crc kubenswrapper[4714]: I0129 16:36:42.476272 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hqc4j/must-gather-684p8" event={"ID":"ff3dd7de-0ec3-4550-ba38-0d67405a2671","Type":"ContainerStarted","Data":"9fd93823c7fddefb604b23f3bb0a407aa212a848b3dc9ccb4d9664567595015f"} Jan 29 16:36:42 crc kubenswrapper[4714]: I0129 16:36:42.476324 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hqc4j/must-gather-684p8" event={"ID":"ff3dd7de-0ec3-4550-ba38-0d67405a2671","Type":"ContainerStarted","Data":"2a5fcee3c64fc545896c1c1e2d77c4828add1a5979b9e718e64ba8cc08bf61fa"} Jan 29 16:36:43 crc kubenswrapper[4714]: I0129 16:36:43.186874 4714 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 16:36:43 crc kubenswrapper[4714]: I0129 16:36:43.486036 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hqc4j/must-gather-684p8" event={"ID":"ff3dd7de-0ec3-4550-ba38-0d67405a2671","Type":"ContainerStarted","Data":"3660194f8ee015e6e72c5a402b3814eb8d75a532bbd31644dbe43ccd17971ac5"} Jan 29 16:36:43 crc kubenswrapper[4714]: I0129 16:36:43.511122 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-hqc4j/must-gather-684p8" podStartSLOduration=2.511101412 podStartE2EDuration="2.511101412s" podCreationTimestamp="2026-01-29 16:36:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 16:36:43.510654589 +0000 UTC m=+1610.031155719" watchObservedRunningTime="2026-01-29 16:36:43.511101412 +0000 UTC m=+1610.031602552" Jan 29 16:36:44 crc kubenswrapper[4714]: I0129 16:36:44.494536 4714 generic.go:334] "Generic (PLEG): container finished" podID="8c12ad14-f878-42a1-a168-bad4026ec2dd" containerID="44fbc675db8c185d9a6165ba1fa89d5622594c14059d2b38df5556992739743d" exitCode=0 Jan 29 16:36:44 crc kubenswrapper[4714]: I0129 16:36:44.494591 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdwq5" event={"ID":"8c12ad14-f878-42a1-a168-bad4026ec2dd","Type":"ContainerDied","Data":"44fbc675db8c185d9a6165ba1fa89d5622594c14059d2b38df5556992739743d"} Jan 29 16:36:45 crc kubenswrapper[4714]: I0129 16:36:45.500697 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdwq5" event={"ID":"8c12ad14-f878-42a1-a168-bad4026ec2dd","Type":"ContainerStarted","Data":"0acf2b33ba53c938c17f8a546ede46efa1c819b32b82b89cd8a55e99bcbea2e7"} Jan 29 16:36:45 crc kubenswrapper[4714]: I0129 16:36:45.533459 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wdwq5" podStartSLOduration=3.455751704 podStartE2EDuration="10m51.533438037s" podCreationTimestamp="2026-01-29 16:25:54 +0000 UTC" firstStartedPulling="2026-01-29 16:25:56.810455116 +0000 UTC m=+963.330956236" lastFinishedPulling="2026-01-29 16:36:44.888141439 +0000 UTC m=+1611.408642569" observedRunningTime="2026-01-29 16:36:45.529292289 +0000 UTC m=+1612.049793419" watchObservedRunningTime="2026-01-29 16:36:45.533438037 +0000 UTC m=+1612.053939157" Jan 29 16:36:55 crc kubenswrapper[4714]: I0129 16:36:55.274017 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wdwq5" Jan 29 16:36:55 crc kubenswrapper[4714]: I0129 16:36:55.274580 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wdwq5" Jan 29 16:36:55 crc kubenswrapper[4714]: I0129 16:36:55.320363 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wdwq5" Jan 29 16:36:55 crc kubenswrapper[4714]: I0129 16:36:55.632429 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wdwq5" Jan 29 16:36:55 crc kubenswrapper[4714]: I0129 16:36:55.683917 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdwq5"] Jan 29 16:36:56 crc kubenswrapper[4714]: I0129 16:36:56.226342 4714 scope.go:117] "RemoveContainer" containerID="be7a968b80d5f3fb2bec436bd6006753f8b129fcc60e1b9be5f43a75c59f2e55" Jan 29 16:36:56 crc kubenswrapper[4714]: I0129 16:36:56.299820 4714 scope.go:117] "RemoveContainer" containerID="5aeecda1a40201485f4391ac8cf0a5c17c26ea2f9167f27b15c6c783da6f0f44" Jan 29 16:36:56 crc kubenswrapper[4714]: I0129 16:36:56.323227 4714 scope.go:117] "RemoveContainer" containerID="ad6e6492e17aa0045196d2d7816583e0511c88fa1ba9566c638560f377a604b8" Jan 29 16:36:56 crc kubenswrapper[4714]: I0129 16:36:56.343586 4714 scope.go:117] "RemoveContainer" containerID="b17d360d1529ce324f10ea0628f0cee292edf28a50e5d75cc7f2606e49a8da6e" Jan 29 16:36:56 crc kubenswrapper[4714]: I0129 16:36:56.367977 4714 scope.go:117] "RemoveContainer" containerID="b09b52cf99e966280d15ceb6a6529b45a9303070f260a1e02acc4b1cf0da02c3" Jan 29 16:36:57 crc kubenswrapper[4714]: I0129 16:36:57.575159 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wdwq5" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" containerName="registry-server" containerID="cri-o://0acf2b33ba53c938c17f8a546ede46efa1c819b32b82b89cd8a55e99bcbea2e7" gracePeriod=2 Jan 29 16:36:57 crc kubenswrapper[4714]: I0129 16:36:57.844186 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:36:57 crc kubenswrapper[4714]: I0129 16:36:57.844237 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.233007 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wdwq5" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.336698 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c12ad14-f878-42a1-a168-bad4026ec2dd-utilities\") pod \"8c12ad14-f878-42a1-a168-bad4026ec2dd\" (UID: \"8c12ad14-f878-42a1-a168-bad4026ec2dd\") " Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.336762 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmm88\" (UniqueName: \"kubernetes.io/projected/8c12ad14-f878-42a1-a168-bad4026ec2dd-kube-api-access-tmm88\") pod \"8c12ad14-f878-42a1-a168-bad4026ec2dd\" (UID: \"8c12ad14-f878-42a1-a168-bad4026ec2dd\") " Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.336836 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c12ad14-f878-42a1-a168-bad4026ec2dd-catalog-content\") pod \"8c12ad14-f878-42a1-a168-bad4026ec2dd\" (UID: \"8c12ad14-f878-42a1-a168-bad4026ec2dd\") " Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.341983 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c12ad14-f878-42a1-a168-bad4026ec2dd-utilities" (OuterVolumeSpecName: "utilities") pod "8c12ad14-f878-42a1-a168-bad4026ec2dd" (UID: "8c12ad14-f878-42a1-a168-bad4026ec2dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.352130 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c12ad14-f878-42a1-a168-bad4026ec2dd-kube-api-access-tmm88" (OuterVolumeSpecName: "kube-api-access-tmm88") pod "8c12ad14-f878-42a1-a168-bad4026ec2dd" (UID: "8c12ad14-f878-42a1-a168-bad4026ec2dd"). InnerVolumeSpecName "kube-api-access-tmm88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.363779 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c12ad14-f878-42a1-a168-bad4026ec2dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c12ad14-f878-42a1-a168-bad4026ec2dd" (UID: "8c12ad14-f878-42a1-a168-bad4026ec2dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.438611 4714 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c12ad14-f878-42a1-a168-bad4026ec2dd-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.438650 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmm88\" (UniqueName: \"kubernetes.io/projected/8c12ad14-f878-42a1-a168-bad4026ec2dd-kube-api-access-tmm88\") on node \"crc\" DevicePath \"\"" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.438663 4714 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c12ad14-f878-42a1-a168-bad4026ec2dd-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.592307 4714 generic.go:334] "Generic (PLEG): container finished" podID="8c12ad14-f878-42a1-a168-bad4026ec2dd" containerID="0acf2b33ba53c938c17f8a546ede46efa1c819b32b82b89cd8a55e99bcbea2e7" exitCode=0 Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.592352 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdwq5" event={"ID":"8c12ad14-f878-42a1-a168-bad4026ec2dd","Type":"ContainerDied","Data":"0acf2b33ba53c938c17f8a546ede46efa1c819b32b82b89cd8a55e99bcbea2e7"} Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.592423 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdwq5" event={"ID":"8c12ad14-f878-42a1-a168-bad4026ec2dd","Type":"ContainerDied","Data":"e9d23b1dec5222eaf00f0f2fac6279153030320fe12205a5c55c774a975165f4"} Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.592447 4714 scope.go:117] "RemoveContainer" containerID="0acf2b33ba53c938c17f8a546ede46efa1c819b32b82b89cd8a55e99bcbea2e7" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.592459 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wdwq5" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.623193 4714 scope.go:117] "RemoveContainer" containerID="44fbc675db8c185d9a6165ba1fa89d5622594c14059d2b38df5556992739743d" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.634496 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdwq5"] Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.644674 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdwq5"] Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.645961 4714 scope.go:117] "RemoveContainer" containerID="efa3dccd677eb8ead685e31efa78ca9b7337f378a73a59452840851ca329f89c" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.676904 4714 scope.go:117] "RemoveContainer" containerID="0acf2b33ba53c938c17f8a546ede46efa1c819b32b82b89cd8a55e99bcbea2e7" Jan 29 16:36:58 crc kubenswrapper[4714]: E0129 16:36:58.677400 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0acf2b33ba53c938c17f8a546ede46efa1c819b32b82b89cd8a55e99bcbea2e7\": container with ID starting with 0acf2b33ba53c938c17f8a546ede46efa1c819b32b82b89cd8a55e99bcbea2e7 not found: ID does not exist" containerID="0acf2b33ba53c938c17f8a546ede46efa1c819b32b82b89cd8a55e99bcbea2e7" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.677439 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0acf2b33ba53c938c17f8a546ede46efa1c819b32b82b89cd8a55e99bcbea2e7"} err="failed to get container status \"0acf2b33ba53c938c17f8a546ede46efa1c819b32b82b89cd8a55e99bcbea2e7\": rpc error: code = NotFound desc = could not find container \"0acf2b33ba53c938c17f8a546ede46efa1c819b32b82b89cd8a55e99bcbea2e7\": container with ID starting with 0acf2b33ba53c938c17f8a546ede46efa1c819b32b82b89cd8a55e99bcbea2e7 not found: ID does not exist" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.677466 4714 scope.go:117] "RemoveContainer" containerID="44fbc675db8c185d9a6165ba1fa89d5622594c14059d2b38df5556992739743d" Jan 29 16:36:58 crc kubenswrapper[4714]: E0129 16:36:58.677865 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44fbc675db8c185d9a6165ba1fa89d5622594c14059d2b38df5556992739743d\": container with ID starting with 44fbc675db8c185d9a6165ba1fa89d5622594c14059d2b38df5556992739743d not found: ID does not exist" containerID="44fbc675db8c185d9a6165ba1fa89d5622594c14059d2b38df5556992739743d" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.677901 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44fbc675db8c185d9a6165ba1fa89d5622594c14059d2b38df5556992739743d"} err="failed to get container status \"44fbc675db8c185d9a6165ba1fa89d5622594c14059d2b38df5556992739743d\": rpc error: code = NotFound desc = could not find container \"44fbc675db8c185d9a6165ba1fa89d5622594c14059d2b38df5556992739743d\": container with ID starting with 44fbc675db8c185d9a6165ba1fa89d5622594c14059d2b38df5556992739743d not found: ID does not exist" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.677921 4714 scope.go:117] "RemoveContainer" containerID="efa3dccd677eb8ead685e31efa78ca9b7337f378a73a59452840851ca329f89c" Jan 29 16:36:58 crc kubenswrapper[4714]: E0129 16:36:58.678207 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efa3dccd677eb8ead685e31efa78ca9b7337f378a73a59452840851ca329f89c\": container with ID starting with efa3dccd677eb8ead685e31efa78ca9b7337f378a73a59452840851ca329f89c not found: ID does not exist" containerID="efa3dccd677eb8ead685e31efa78ca9b7337f378a73a59452840851ca329f89c" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.678234 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efa3dccd677eb8ead685e31efa78ca9b7337f378a73a59452840851ca329f89c"} err="failed to get container status \"efa3dccd677eb8ead685e31efa78ca9b7337f378a73a59452840851ca329f89c\": rpc error: code = NotFound desc = could not find container \"efa3dccd677eb8ead685e31efa78ca9b7337f378a73a59452840851ca329f89c\": container with ID starting with efa3dccd677eb8ead685e31efa78ca9b7337f378a73a59452840851ca329f89c not found: ID does not exist" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.963479 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2lqb9"] Jan 29 16:36:58 crc kubenswrapper[4714]: E0129 16:36:58.963805 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" containerName="registry-server" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.963833 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" containerName="registry-server" Jan 29 16:36:58 crc kubenswrapper[4714]: E0129 16:36:58.963858 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" containerName="extract-content" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.963870 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" containerName="extract-content" Jan 29 16:36:58 crc kubenswrapper[4714]: E0129 16:36:58.963891 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" containerName="extract-utilities" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.963905 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" containerName="extract-utilities" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.967177 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" containerName="registry-server" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.968554 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2lqb9" Jan 29 16:36:58 crc kubenswrapper[4714]: I0129 16:36:58.975566 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2lqb9"] Jan 29 16:36:59 crc kubenswrapper[4714]: I0129 16:36:59.046803 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b6f14e1-18de-4e4f-852e-caadbfd35b71-catalog-content\") pod \"redhat-marketplace-2lqb9\" (UID: \"5b6f14e1-18de-4e4f-852e-caadbfd35b71\") " pod="openshift-marketplace/redhat-marketplace-2lqb9" Jan 29 16:36:59 crc kubenswrapper[4714]: I0129 16:36:59.047374 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b6f14e1-18de-4e4f-852e-caadbfd35b71-utilities\") pod \"redhat-marketplace-2lqb9\" (UID: \"5b6f14e1-18de-4e4f-852e-caadbfd35b71\") " pod="openshift-marketplace/redhat-marketplace-2lqb9" Jan 29 16:36:59 crc kubenswrapper[4714]: I0129 16:36:59.047602 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlgsv\" (UniqueName: \"kubernetes.io/projected/5b6f14e1-18de-4e4f-852e-caadbfd35b71-kube-api-access-dlgsv\") pod \"redhat-marketplace-2lqb9\" (UID: \"5b6f14e1-18de-4e4f-852e-caadbfd35b71\") " pod="openshift-marketplace/redhat-marketplace-2lqb9" Jan 29 16:36:59 crc kubenswrapper[4714]: I0129 16:36:59.148765 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlgsv\" (UniqueName: \"kubernetes.io/projected/5b6f14e1-18de-4e4f-852e-caadbfd35b71-kube-api-access-dlgsv\") pod \"redhat-marketplace-2lqb9\" (UID: \"5b6f14e1-18de-4e4f-852e-caadbfd35b71\") " pod="openshift-marketplace/redhat-marketplace-2lqb9" Jan 29 16:36:59 crc kubenswrapper[4714]: I0129 16:36:59.149036 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b6f14e1-18de-4e4f-852e-caadbfd35b71-catalog-content\") pod \"redhat-marketplace-2lqb9\" (UID: \"5b6f14e1-18de-4e4f-852e-caadbfd35b71\") " pod="openshift-marketplace/redhat-marketplace-2lqb9" Jan 29 16:36:59 crc kubenswrapper[4714]: I0129 16:36:59.149378 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b6f14e1-18de-4e4f-852e-caadbfd35b71-utilities\") pod \"redhat-marketplace-2lqb9\" (UID: \"5b6f14e1-18de-4e4f-852e-caadbfd35b71\") " pod="openshift-marketplace/redhat-marketplace-2lqb9" Jan 29 16:36:59 crc kubenswrapper[4714]: I0129 16:36:59.149845 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b6f14e1-18de-4e4f-852e-caadbfd35b71-utilities\") pod \"redhat-marketplace-2lqb9\" (UID: \"5b6f14e1-18de-4e4f-852e-caadbfd35b71\") " pod="openshift-marketplace/redhat-marketplace-2lqb9" Jan 29 16:36:59 crc kubenswrapper[4714]: I0129 16:36:59.149851 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b6f14e1-18de-4e4f-852e-caadbfd35b71-catalog-content\") pod \"redhat-marketplace-2lqb9\" (UID: \"5b6f14e1-18de-4e4f-852e-caadbfd35b71\") " pod="openshift-marketplace/redhat-marketplace-2lqb9" Jan 29 16:36:59 crc kubenswrapper[4714]: I0129 16:36:59.174833 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlgsv\" (UniqueName: \"kubernetes.io/projected/5b6f14e1-18de-4e4f-852e-caadbfd35b71-kube-api-access-dlgsv\") pod \"redhat-marketplace-2lqb9\" (UID: \"5b6f14e1-18de-4e4f-852e-caadbfd35b71\") " pod="openshift-marketplace/redhat-marketplace-2lqb9" Jan 29 16:36:59 crc kubenswrapper[4714]: I0129 16:36:59.286120 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2lqb9" Jan 29 16:36:59 crc kubenswrapper[4714]: I0129 16:36:59.510541 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2lqb9"] Jan 29 16:36:59 crc kubenswrapper[4714]: W0129 16:36:59.515794 4714 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b6f14e1_18de_4e4f_852e_caadbfd35b71.slice/crio-f02d4ba09eb3f5d2bb88a68856dd6a54f7621db0ebd026df1821a9b0f7c172cf WatchSource:0}: Error finding container f02d4ba09eb3f5d2bb88a68856dd6a54f7621db0ebd026df1821a9b0f7c172cf: Status 404 returned error can't find the container with id f02d4ba09eb3f5d2bb88a68856dd6a54f7621db0ebd026df1821a9b0f7c172cf Jan 29 16:36:59 crc kubenswrapper[4714]: I0129 16:36:59.597283 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2lqb9" event={"ID":"5b6f14e1-18de-4e4f-852e-caadbfd35b71","Type":"ContainerStarted","Data":"f02d4ba09eb3f5d2bb88a68856dd6a54f7621db0ebd026df1821a9b0f7c172cf"} Jan 29 16:37:00 crc kubenswrapper[4714]: I0129 16:37:00.195449 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c12ad14-f878-42a1-a168-bad4026ec2dd" path="/var/lib/kubelet/pods/8c12ad14-f878-42a1-a168-bad4026ec2dd/volumes" Jan 29 16:37:00 crc kubenswrapper[4714]: I0129 16:37:00.610731 4714 generic.go:334] "Generic (PLEG): container finished" podID="5b6f14e1-18de-4e4f-852e-caadbfd35b71" containerID="2f13fd962378375cad6ff392c83e16f142844861e33517b57b33783403f85fe4" exitCode=0 Jan 29 16:37:00 crc kubenswrapper[4714]: I0129 16:37:00.610838 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2lqb9" event={"ID":"5b6f14e1-18de-4e4f-852e-caadbfd35b71","Type":"ContainerDied","Data":"2f13fd962378375cad6ff392c83e16f142844861e33517b57b33783403f85fe4"} Jan 29 16:37:01 crc kubenswrapper[4714]: I0129 16:37:01.620848 4714 generic.go:334] "Generic (PLEG): container finished" podID="5b6f14e1-18de-4e4f-852e-caadbfd35b71" containerID="eff5c7821884beca01fac4daf91eae898a6dd4fe1c53607f72659ececca4c316" exitCode=0 Jan 29 16:37:01 crc kubenswrapper[4714]: I0129 16:37:01.620921 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2lqb9" event={"ID":"5b6f14e1-18de-4e4f-852e-caadbfd35b71","Type":"ContainerDied","Data":"eff5c7821884beca01fac4daf91eae898a6dd4fe1c53607f72659ececca4c316"} Jan 29 16:37:02 crc kubenswrapper[4714]: I0129 16:37:02.629726 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2lqb9" event={"ID":"5b6f14e1-18de-4e4f-852e-caadbfd35b71","Type":"ContainerStarted","Data":"d368dec2ee8c5ee9383ebfac5d4bd159dc5ec20a2bce8b644dccca52e523a8b8"} Jan 29 16:37:07 crc kubenswrapper[4714]: I0129 16:37:07.851511 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2lqb9" podStartSLOduration=8.424355507 podStartE2EDuration="9.85146758s" podCreationTimestamp="2026-01-29 16:36:58 +0000 UTC" firstStartedPulling="2026-01-29 16:37:00.613751411 +0000 UTC m=+1627.134252561" lastFinishedPulling="2026-01-29 16:37:02.040863504 +0000 UTC m=+1628.561364634" observedRunningTime="2026-01-29 16:37:02.663416867 +0000 UTC m=+1629.183917997" watchObservedRunningTime="2026-01-29 16:37:07.85146758 +0000 UTC m=+1634.371968730" Jan 29 16:37:07 crc kubenswrapper[4714]: I0129 16:37:07.854567 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jhx4h"] Jan 29 16:37:07 crc kubenswrapper[4714]: I0129 16:37:07.856854 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jhx4h" Jan 29 16:37:07 crc kubenswrapper[4714]: I0129 16:37:07.860770 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jhx4h"] Jan 29 16:37:07 crc kubenswrapper[4714]: I0129 16:37:07.969885 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f29ce50e-bbd2-4fe5-a032-5163f3f80ca0-utilities\") pod \"community-operators-jhx4h\" (UID: \"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0\") " pod="openshift-marketplace/community-operators-jhx4h" Jan 29 16:37:07 crc kubenswrapper[4714]: I0129 16:37:07.970009 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hctqd\" (UniqueName: \"kubernetes.io/projected/f29ce50e-bbd2-4fe5-a032-5163f3f80ca0-kube-api-access-hctqd\") pod \"community-operators-jhx4h\" (UID: \"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0\") " pod="openshift-marketplace/community-operators-jhx4h" Jan 29 16:37:07 crc kubenswrapper[4714]: I0129 16:37:07.970149 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f29ce50e-bbd2-4fe5-a032-5163f3f80ca0-catalog-content\") pod \"community-operators-jhx4h\" (UID: \"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0\") " pod="openshift-marketplace/community-operators-jhx4h" Jan 29 16:37:08 crc kubenswrapper[4714]: I0129 16:37:08.071855 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f29ce50e-bbd2-4fe5-a032-5163f3f80ca0-catalog-content\") pod \"community-operators-jhx4h\" (UID: \"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0\") " pod="openshift-marketplace/community-operators-jhx4h" Jan 29 16:37:08 crc kubenswrapper[4714]: I0129 16:37:08.071911 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f29ce50e-bbd2-4fe5-a032-5163f3f80ca0-utilities\") pod \"community-operators-jhx4h\" (UID: \"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0\") " pod="openshift-marketplace/community-operators-jhx4h" Jan 29 16:37:08 crc kubenswrapper[4714]: I0129 16:37:08.071970 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hctqd\" (UniqueName: \"kubernetes.io/projected/f29ce50e-bbd2-4fe5-a032-5163f3f80ca0-kube-api-access-hctqd\") pod \"community-operators-jhx4h\" (UID: \"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0\") " pod="openshift-marketplace/community-operators-jhx4h" Jan 29 16:37:08 crc kubenswrapper[4714]: I0129 16:37:08.072430 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f29ce50e-bbd2-4fe5-a032-5163f3f80ca0-catalog-content\") pod \"community-operators-jhx4h\" (UID: \"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0\") " pod="openshift-marketplace/community-operators-jhx4h" Jan 29 16:37:08 crc kubenswrapper[4714]: I0129 16:37:08.072483 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f29ce50e-bbd2-4fe5-a032-5163f3f80ca0-utilities\") pod \"community-operators-jhx4h\" (UID: \"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0\") " pod="openshift-marketplace/community-operators-jhx4h" Jan 29 16:37:08 crc kubenswrapper[4714]: I0129 16:37:08.102556 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hctqd\" (UniqueName: \"kubernetes.io/projected/f29ce50e-bbd2-4fe5-a032-5163f3f80ca0-kube-api-access-hctqd\") pod \"community-operators-jhx4h\" (UID: \"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0\") " pod="openshift-marketplace/community-operators-jhx4h" Jan 29 16:37:08 crc kubenswrapper[4714]: I0129 16:37:08.175180 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jhx4h" Jan 29 16:37:08 crc kubenswrapper[4714]: I0129 16:37:08.464624 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jhx4h"] Jan 29 16:37:08 crc kubenswrapper[4714]: I0129 16:37:08.667737 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhx4h" event={"ID":"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0","Type":"ContainerStarted","Data":"bb661581f20750aeeeb2dcee345a2fe34987995b5a11b5fadc693c52078fdfde"} Jan 29 16:37:09 crc kubenswrapper[4714]: I0129 16:37:09.286814 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2lqb9" Jan 29 16:37:09 crc kubenswrapper[4714]: I0129 16:37:09.286960 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2lqb9" Jan 29 16:37:09 crc kubenswrapper[4714]: I0129 16:37:09.343602 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2lqb9" Jan 29 16:37:09 crc kubenswrapper[4714]: I0129 16:37:09.680969 4714 generic.go:334] "Generic (PLEG): container finished" podID="f29ce50e-bbd2-4fe5-a032-5163f3f80ca0" containerID="12528aa63a645da73775930ec51922ba0e34d726bc9e8bf443beb9cd97abf1ae" exitCode=0 Jan 29 16:37:09 crc kubenswrapper[4714]: I0129 16:37:09.681052 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhx4h" event={"ID":"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0","Type":"ContainerDied","Data":"12528aa63a645da73775930ec51922ba0e34d726bc9e8bf443beb9cd97abf1ae"} Jan 29 16:37:09 crc kubenswrapper[4714]: I0129 16:37:09.729149 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2lqb9" Jan 29 16:37:10 crc kubenswrapper[4714]: I0129 16:37:10.688348 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhx4h" event={"ID":"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0","Type":"ContainerStarted","Data":"70f1c853b2f45447d864ab1780c740cfb8b5e66e1b37ea10c838d033de579edc"} Jan 29 16:37:11 crc kubenswrapper[4714]: I0129 16:37:11.629667 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2lqb9"] Jan 29 16:37:11 crc kubenswrapper[4714]: I0129 16:37:11.693363 4714 generic.go:334] "Generic (PLEG): container finished" podID="f29ce50e-bbd2-4fe5-a032-5163f3f80ca0" containerID="70f1c853b2f45447d864ab1780c740cfb8b5e66e1b37ea10c838d033de579edc" exitCode=0 Jan 29 16:37:11 crc kubenswrapper[4714]: I0129 16:37:11.693421 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhx4h" event={"ID":"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0","Type":"ContainerDied","Data":"70f1c853b2f45447d864ab1780c740cfb8b5e66e1b37ea10c838d033de579edc"} Jan 29 16:37:12 crc kubenswrapper[4714]: I0129 16:37:12.700051 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhx4h" event={"ID":"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0","Type":"ContainerStarted","Data":"d8f87ee4c8ffa275df436d8576ba01192465ef13d6a102cdef3b3eec7691fb41"} Jan 29 16:37:12 crc kubenswrapper[4714]: I0129 16:37:12.700172 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2lqb9" podUID="5b6f14e1-18de-4e4f-852e-caadbfd35b71" containerName="registry-server" containerID="cri-o://d368dec2ee8c5ee9383ebfac5d4bd159dc5ec20a2bce8b644dccca52e523a8b8" gracePeriod=2 Jan 29 16:37:12 crc kubenswrapper[4714]: I0129 16:37:12.724906 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jhx4h" podStartSLOduration=3.299617414 podStartE2EDuration="5.724885234s" podCreationTimestamp="2026-01-29 16:37:07 +0000 UTC" firstStartedPulling="2026-01-29 16:37:09.683130943 +0000 UTC m=+1636.203632073" lastFinishedPulling="2026-01-29 16:37:12.108398773 +0000 UTC m=+1638.628899893" observedRunningTime="2026-01-29 16:37:12.720922341 +0000 UTC m=+1639.241423461" watchObservedRunningTime="2026-01-29 16:37:12.724885234 +0000 UTC m=+1639.245386364" Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.041873 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2lqb9" Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.150233 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b6f14e1-18de-4e4f-852e-caadbfd35b71-utilities\") pod \"5b6f14e1-18de-4e4f-852e-caadbfd35b71\" (UID: \"5b6f14e1-18de-4e4f-852e-caadbfd35b71\") " Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.150277 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlgsv\" (UniqueName: \"kubernetes.io/projected/5b6f14e1-18de-4e4f-852e-caadbfd35b71-kube-api-access-dlgsv\") pod \"5b6f14e1-18de-4e4f-852e-caadbfd35b71\" (UID: \"5b6f14e1-18de-4e4f-852e-caadbfd35b71\") " Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.150329 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b6f14e1-18de-4e4f-852e-caadbfd35b71-catalog-content\") pod \"5b6f14e1-18de-4e4f-852e-caadbfd35b71\" (UID: \"5b6f14e1-18de-4e4f-852e-caadbfd35b71\") " Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.151003 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b6f14e1-18de-4e4f-852e-caadbfd35b71-utilities" (OuterVolumeSpecName: "utilities") pod "5b6f14e1-18de-4e4f-852e-caadbfd35b71" (UID: "5b6f14e1-18de-4e4f-852e-caadbfd35b71"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.164568 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b6f14e1-18de-4e4f-852e-caadbfd35b71-kube-api-access-dlgsv" (OuterVolumeSpecName: "kube-api-access-dlgsv") pod "5b6f14e1-18de-4e4f-852e-caadbfd35b71" (UID: "5b6f14e1-18de-4e4f-852e-caadbfd35b71"). InnerVolumeSpecName "kube-api-access-dlgsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.181216 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b6f14e1-18de-4e4f-852e-caadbfd35b71-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b6f14e1-18de-4e4f-852e-caadbfd35b71" (UID: "5b6f14e1-18de-4e4f-852e-caadbfd35b71"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.251153 4714 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b6f14e1-18de-4e4f-852e-caadbfd35b71-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.251185 4714 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b6f14e1-18de-4e4f-852e-caadbfd35b71-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.251194 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlgsv\" (UniqueName: \"kubernetes.io/projected/5b6f14e1-18de-4e4f-852e-caadbfd35b71-kube-api-access-dlgsv\") on node \"crc\" DevicePath \"\"" Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.711287 4714 generic.go:334] "Generic (PLEG): container finished" podID="5b6f14e1-18de-4e4f-852e-caadbfd35b71" containerID="d368dec2ee8c5ee9383ebfac5d4bd159dc5ec20a2bce8b644dccca52e523a8b8" exitCode=0 Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.711316 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2lqb9" event={"ID":"5b6f14e1-18de-4e4f-852e-caadbfd35b71","Type":"ContainerDied","Data":"d368dec2ee8c5ee9383ebfac5d4bd159dc5ec20a2bce8b644dccca52e523a8b8"} Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.711628 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2lqb9" event={"ID":"5b6f14e1-18de-4e4f-852e-caadbfd35b71","Type":"ContainerDied","Data":"f02d4ba09eb3f5d2bb88a68856dd6a54f7621db0ebd026df1821a9b0f7c172cf"} Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.711660 4714 scope.go:117] "RemoveContainer" containerID="d368dec2ee8c5ee9383ebfac5d4bd159dc5ec20a2bce8b644dccca52e523a8b8" Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.711377 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2lqb9" Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.779624 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2lqb9"] Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.800606 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2lqb9"] Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.803981 4714 scope.go:117] "RemoveContainer" containerID="eff5c7821884beca01fac4daf91eae898a6dd4fe1c53607f72659ececca4c316" Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.857264 4714 scope.go:117] "RemoveContainer" containerID="2f13fd962378375cad6ff392c83e16f142844861e33517b57b33783403f85fe4" Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.881743 4714 scope.go:117] "RemoveContainer" containerID="d368dec2ee8c5ee9383ebfac5d4bd159dc5ec20a2bce8b644dccca52e523a8b8" Jan 29 16:37:13 crc kubenswrapper[4714]: E0129 16:37:13.882335 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d368dec2ee8c5ee9383ebfac5d4bd159dc5ec20a2bce8b644dccca52e523a8b8\": container with ID starting with d368dec2ee8c5ee9383ebfac5d4bd159dc5ec20a2bce8b644dccca52e523a8b8 not found: ID does not exist" containerID="d368dec2ee8c5ee9383ebfac5d4bd159dc5ec20a2bce8b644dccca52e523a8b8" Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.882400 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d368dec2ee8c5ee9383ebfac5d4bd159dc5ec20a2bce8b644dccca52e523a8b8"} err="failed to get container status \"d368dec2ee8c5ee9383ebfac5d4bd159dc5ec20a2bce8b644dccca52e523a8b8\": rpc error: code = NotFound desc = could not find container \"d368dec2ee8c5ee9383ebfac5d4bd159dc5ec20a2bce8b644dccca52e523a8b8\": container with ID starting with d368dec2ee8c5ee9383ebfac5d4bd159dc5ec20a2bce8b644dccca52e523a8b8 not found: ID does not exist" Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.882444 4714 scope.go:117] "RemoveContainer" containerID="eff5c7821884beca01fac4daf91eae898a6dd4fe1c53607f72659ececca4c316" Jan 29 16:37:13 crc kubenswrapper[4714]: E0129 16:37:13.882815 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eff5c7821884beca01fac4daf91eae898a6dd4fe1c53607f72659ececca4c316\": container with ID starting with eff5c7821884beca01fac4daf91eae898a6dd4fe1c53607f72659ececca4c316 not found: ID does not exist" containerID="eff5c7821884beca01fac4daf91eae898a6dd4fe1c53607f72659ececca4c316" Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.882860 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eff5c7821884beca01fac4daf91eae898a6dd4fe1c53607f72659ececca4c316"} err="failed to get container status \"eff5c7821884beca01fac4daf91eae898a6dd4fe1c53607f72659ececca4c316\": rpc error: code = NotFound desc = could not find container \"eff5c7821884beca01fac4daf91eae898a6dd4fe1c53607f72659ececca4c316\": container with ID starting with eff5c7821884beca01fac4daf91eae898a6dd4fe1c53607f72659ececca4c316 not found: ID does not exist" Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.882900 4714 scope.go:117] "RemoveContainer" containerID="2f13fd962378375cad6ff392c83e16f142844861e33517b57b33783403f85fe4" Jan 29 16:37:13 crc kubenswrapper[4714]: E0129 16:37:13.883692 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f13fd962378375cad6ff392c83e16f142844861e33517b57b33783403f85fe4\": container with ID starting with 2f13fd962378375cad6ff392c83e16f142844861e33517b57b33783403f85fe4 not found: ID does not exist" containerID="2f13fd962378375cad6ff392c83e16f142844861e33517b57b33783403f85fe4" Jan 29 16:37:13 crc kubenswrapper[4714]: I0129 16:37:13.883752 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f13fd962378375cad6ff392c83e16f142844861e33517b57b33783403f85fe4"} err="failed to get container status \"2f13fd962378375cad6ff392c83e16f142844861e33517b57b33783403f85fe4\": rpc error: code = NotFound desc = could not find container \"2f13fd962378375cad6ff392c83e16f142844861e33517b57b33783403f85fe4\": container with ID starting with 2f13fd962378375cad6ff392c83e16f142844861e33517b57b33783403f85fe4 not found: ID does not exist" Jan 29 16:37:14 crc kubenswrapper[4714]: I0129 16:37:14.196868 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b6f14e1-18de-4e4f-852e-caadbfd35b71" path="/var/lib/kubelet/pods/5b6f14e1-18de-4e4f-852e-caadbfd35b71/volumes" Jan 29 16:37:17 crc kubenswrapper[4714]: I0129 16:37:17.029784 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qdbvx"] Jan 29 16:37:17 crc kubenswrapper[4714]: E0129 16:37:17.030385 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b6f14e1-18de-4e4f-852e-caadbfd35b71" containerName="registry-server" Jan 29 16:37:17 crc kubenswrapper[4714]: I0129 16:37:17.030401 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b6f14e1-18de-4e4f-852e-caadbfd35b71" containerName="registry-server" Jan 29 16:37:17 crc kubenswrapper[4714]: E0129 16:37:17.030417 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b6f14e1-18de-4e4f-852e-caadbfd35b71" containerName="extract-utilities" Jan 29 16:37:17 crc kubenswrapper[4714]: I0129 16:37:17.030425 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b6f14e1-18de-4e4f-852e-caadbfd35b71" containerName="extract-utilities" Jan 29 16:37:17 crc kubenswrapper[4714]: E0129 16:37:17.030446 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b6f14e1-18de-4e4f-852e-caadbfd35b71" containerName="extract-content" Jan 29 16:37:17 crc kubenswrapper[4714]: I0129 16:37:17.030454 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b6f14e1-18de-4e4f-852e-caadbfd35b71" containerName="extract-content" Jan 29 16:37:17 crc kubenswrapper[4714]: I0129 16:37:17.030589 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b6f14e1-18de-4e4f-852e-caadbfd35b71" containerName="registry-server" Jan 29 16:37:17 crc kubenswrapper[4714]: I0129 16:37:17.031650 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qdbvx" Jan 29 16:37:17 crc kubenswrapper[4714]: I0129 16:37:17.053398 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qdbvx"] Jan 29 16:37:17 crc kubenswrapper[4714]: I0129 16:37:17.100476 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e6b3552-9d7a-459f-a4c4-67da40c67c15-catalog-content\") pod \"certified-operators-qdbvx\" (UID: \"4e6b3552-9d7a-459f-a4c4-67da40c67c15\") " pod="openshift-marketplace/certified-operators-qdbvx" Jan 29 16:37:17 crc kubenswrapper[4714]: I0129 16:37:17.100524 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e6b3552-9d7a-459f-a4c4-67da40c67c15-utilities\") pod \"certified-operators-qdbvx\" (UID: \"4e6b3552-9d7a-459f-a4c4-67da40c67c15\") " pod="openshift-marketplace/certified-operators-qdbvx" Jan 29 16:37:17 crc kubenswrapper[4714]: I0129 16:37:17.100558 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7b4g\" (UniqueName: \"kubernetes.io/projected/4e6b3552-9d7a-459f-a4c4-67da40c67c15-kube-api-access-j7b4g\") pod \"certified-operators-qdbvx\" (UID: \"4e6b3552-9d7a-459f-a4c4-67da40c67c15\") " pod="openshift-marketplace/certified-operators-qdbvx" Jan 29 16:37:17 crc kubenswrapper[4714]: I0129 16:37:17.202462 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7b4g\" (UniqueName: \"kubernetes.io/projected/4e6b3552-9d7a-459f-a4c4-67da40c67c15-kube-api-access-j7b4g\") pod \"certified-operators-qdbvx\" (UID: \"4e6b3552-9d7a-459f-a4c4-67da40c67c15\") " pod="openshift-marketplace/certified-operators-qdbvx" Jan 29 16:37:17 crc kubenswrapper[4714]: I0129 16:37:17.202652 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e6b3552-9d7a-459f-a4c4-67da40c67c15-catalog-content\") pod \"certified-operators-qdbvx\" (UID: \"4e6b3552-9d7a-459f-a4c4-67da40c67c15\") " pod="openshift-marketplace/certified-operators-qdbvx" Jan 29 16:37:17 crc kubenswrapper[4714]: I0129 16:37:17.202693 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e6b3552-9d7a-459f-a4c4-67da40c67c15-utilities\") pod \"certified-operators-qdbvx\" (UID: \"4e6b3552-9d7a-459f-a4c4-67da40c67c15\") " pod="openshift-marketplace/certified-operators-qdbvx" Jan 29 16:37:17 crc kubenswrapper[4714]: I0129 16:37:17.203246 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e6b3552-9d7a-459f-a4c4-67da40c67c15-utilities\") pod \"certified-operators-qdbvx\" (UID: \"4e6b3552-9d7a-459f-a4c4-67da40c67c15\") " pod="openshift-marketplace/certified-operators-qdbvx" Jan 29 16:37:17 crc kubenswrapper[4714]: I0129 16:37:17.203302 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e6b3552-9d7a-459f-a4c4-67da40c67c15-catalog-content\") pod \"certified-operators-qdbvx\" (UID: \"4e6b3552-9d7a-459f-a4c4-67da40c67c15\") " pod="openshift-marketplace/certified-operators-qdbvx" Jan 29 16:37:17 crc kubenswrapper[4714]: I0129 16:37:17.251013 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7b4g\" (UniqueName: \"kubernetes.io/projected/4e6b3552-9d7a-459f-a4c4-67da40c67c15-kube-api-access-j7b4g\") pod \"certified-operators-qdbvx\" (UID: \"4e6b3552-9d7a-459f-a4c4-67da40c67c15\") " pod="openshift-marketplace/certified-operators-qdbvx" Jan 29 16:37:17 crc kubenswrapper[4714]: I0129 16:37:17.355510 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qdbvx" Jan 29 16:37:17 crc kubenswrapper[4714]: I0129 16:37:17.586781 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qdbvx"] Jan 29 16:37:17 crc kubenswrapper[4714]: I0129 16:37:17.737553 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qdbvx" event={"ID":"4e6b3552-9d7a-459f-a4c4-67da40c67c15","Type":"ContainerStarted","Data":"7cbae5276fb8d757c39c5f0cacfa9727ddbf2eb3306d3df09a254f9f755544f8"} Jan 29 16:37:18 crc kubenswrapper[4714]: I0129 16:37:18.175767 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jhx4h" Jan 29 16:37:18 crc kubenswrapper[4714]: I0129 16:37:18.175856 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jhx4h" Jan 29 16:37:18 crc kubenswrapper[4714]: I0129 16:37:18.231831 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jhx4h" Jan 29 16:37:18 crc kubenswrapper[4714]: I0129 16:37:18.751156 4714 generic.go:334] "Generic (PLEG): container finished" podID="4e6b3552-9d7a-459f-a4c4-67da40c67c15" containerID="654e4e9ffb535db9fca5c14eda678189ef4ed63a10878793979a5fe56e16938d" exitCode=0 Jan 29 16:37:18 crc kubenswrapper[4714]: I0129 16:37:18.751244 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qdbvx" event={"ID":"4e6b3552-9d7a-459f-a4c4-67da40c67c15","Type":"ContainerDied","Data":"654e4e9ffb535db9fca5c14eda678189ef4ed63a10878793979a5fe56e16938d"} Jan 29 16:37:18 crc kubenswrapper[4714]: I0129 16:37:18.808290 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jhx4h" Jan 29 16:37:20 crc kubenswrapper[4714]: I0129 16:37:20.762304 4714 generic.go:334] "Generic (PLEG): container finished" podID="4e6b3552-9d7a-459f-a4c4-67da40c67c15" containerID="cd702a907a78fab939fb238809b451a3868bf29847022a3fd0a1279c986cdaca" exitCode=0 Jan 29 16:37:20 crc kubenswrapper[4714]: I0129 16:37:20.762348 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qdbvx" event={"ID":"4e6b3552-9d7a-459f-a4c4-67da40c67c15","Type":"ContainerDied","Data":"cd702a907a78fab939fb238809b451a3868bf29847022a3fd0a1279c986cdaca"} Jan 29 16:37:21 crc kubenswrapper[4714]: I0129 16:37:21.423956 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jhx4h"] Jan 29 16:37:21 crc kubenswrapper[4714]: I0129 16:37:21.424214 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jhx4h" podUID="f29ce50e-bbd2-4fe5-a032-5163f3f80ca0" containerName="registry-server" containerID="cri-o://d8f87ee4c8ffa275df436d8576ba01192465ef13d6a102cdef3b3eec7691fb41" gracePeriod=2 Jan 29 16:37:21 crc kubenswrapper[4714]: I0129 16:37:21.769557 4714 generic.go:334] "Generic (PLEG): container finished" podID="f29ce50e-bbd2-4fe5-a032-5163f3f80ca0" containerID="d8f87ee4c8ffa275df436d8576ba01192465ef13d6a102cdef3b3eec7691fb41" exitCode=0 Jan 29 16:37:21 crc kubenswrapper[4714]: I0129 16:37:21.769775 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhx4h" event={"ID":"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0","Type":"ContainerDied","Data":"d8f87ee4c8ffa275df436d8576ba01192465ef13d6a102cdef3b3eec7691fb41"} Jan 29 16:37:21 crc kubenswrapper[4714]: I0129 16:37:21.772653 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qdbvx" event={"ID":"4e6b3552-9d7a-459f-a4c4-67da40c67c15","Type":"ContainerStarted","Data":"9632c9e5f7fc6621d6e14cceb0240050190e925aa5472c826d55f03bda4cac18"} Jan 29 16:37:21 crc kubenswrapper[4714]: I0129 16:37:21.797759 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qdbvx" podStartSLOduration=2.27724702 podStartE2EDuration="4.797742406s" podCreationTimestamp="2026-01-29 16:37:17 +0000 UTC" firstStartedPulling="2026-01-29 16:37:18.753113572 +0000 UTC m=+1645.273614732" lastFinishedPulling="2026-01-29 16:37:21.273608998 +0000 UTC m=+1647.794110118" observedRunningTime="2026-01-29 16:37:21.795574334 +0000 UTC m=+1648.316075464" watchObservedRunningTime="2026-01-29 16:37:21.797742406 +0000 UTC m=+1648.318243526" Jan 29 16:37:22 crc kubenswrapper[4714]: I0129 16:37:22.065997 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jhx4h" Jan 29 16:37:22 crc kubenswrapper[4714]: I0129 16:37:22.164919 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f29ce50e-bbd2-4fe5-a032-5163f3f80ca0-utilities\") pod \"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0\" (UID: \"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0\") " Jan 29 16:37:22 crc kubenswrapper[4714]: I0129 16:37:22.165046 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hctqd\" (UniqueName: \"kubernetes.io/projected/f29ce50e-bbd2-4fe5-a032-5163f3f80ca0-kube-api-access-hctqd\") pod \"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0\" (UID: \"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0\") " Jan 29 16:37:22 crc kubenswrapper[4714]: I0129 16:37:22.165121 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f29ce50e-bbd2-4fe5-a032-5163f3f80ca0-catalog-content\") pod \"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0\" (UID: \"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0\") " Jan 29 16:37:22 crc kubenswrapper[4714]: I0129 16:37:22.165578 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f29ce50e-bbd2-4fe5-a032-5163f3f80ca0-utilities" (OuterVolumeSpecName: "utilities") pod "f29ce50e-bbd2-4fe5-a032-5163f3f80ca0" (UID: "f29ce50e-bbd2-4fe5-a032-5163f3f80ca0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:37:22 crc kubenswrapper[4714]: I0129 16:37:22.188240 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f29ce50e-bbd2-4fe5-a032-5163f3f80ca0-kube-api-access-hctqd" (OuterVolumeSpecName: "kube-api-access-hctqd") pod "f29ce50e-bbd2-4fe5-a032-5163f3f80ca0" (UID: "f29ce50e-bbd2-4fe5-a032-5163f3f80ca0"). InnerVolumeSpecName "kube-api-access-hctqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:37:22 crc kubenswrapper[4714]: I0129 16:37:22.267638 4714 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f29ce50e-bbd2-4fe5-a032-5163f3f80ca0-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:37:22 crc kubenswrapper[4714]: I0129 16:37:22.267699 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hctqd\" (UniqueName: \"kubernetes.io/projected/f29ce50e-bbd2-4fe5-a032-5163f3f80ca0-kube-api-access-hctqd\") on node \"crc\" DevicePath \"\"" Jan 29 16:37:22 crc kubenswrapper[4714]: I0129 16:37:22.689117 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f29ce50e-bbd2-4fe5-a032-5163f3f80ca0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f29ce50e-bbd2-4fe5-a032-5163f3f80ca0" (UID: "f29ce50e-bbd2-4fe5-a032-5163f3f80ca0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:37:22 crc kubenswrapper[4714]: I0129 16:37:22.774122 4714 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f29ce50e-bbd2-4fe5-a032-5163f3f80ca0-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:37:22 crc kubenswrapper[4714]: I0129 16:37:22.780200 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jhx4h" Jan 29 16:37:22 crc kubenswrapper[4714]: I0129 16:37:22.780189 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhx4h" event={"ID":"f29ce50e-bbd2-4fe5-a032-5163f3f80ca0","Type":"ContainerDied","Data":"bb661581f20750aeeeb2dcee345a2fe34987995b5a11b5fadc693c52078fdfde"} Jan 29 16:37:22 crc kubenswrapper[4714]: I0129 16:37:22.780272 4714 scope.go:117] "RemoveContainer" containerID="d8f87ee4c8ffa275df436d8576ba01192465ef13d6a102cdef3b3eec7691fb41" Jan 29 16:37:22 crc kubenswrapper[4714]: I0129 16:37:22.800313 4714 scope.go:117] "RemoveContainer" containerID="70f1c853b2f45447d864ab1780c740cfb8b5e66e1b37ea10c838d033de579edc" Jan 29 16:37:22 crc kubenswrapper[4714]: I0129 16:37:22.813514 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jhx4h"] Jan 29 16:37:22 crc kubenswrapper[4714]: I0129 16:37:22.818803 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jhx4h"] Jan 29 16:37:22 crc kubenswrapper[4714]: I0129 16:37:22.829641 4714 scope.go:117] "RemoveContainer" containerID="12528aa63a645da73775930ec51922ba0e34d726bc9e8bf443beb9cd97abf1ae" Jan 29 16:37:24 crc kubenswrapper[4714]: I0129 16:37:24.195327 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f29ce50e-bbd2-4fe5-a032-5163f3f80ca0" path="/var/lib/kubelet/pods/f29ce50e-bbd2-4fe5-a032-5163f3f80ca0/volumes" Jan 29 16:37:27 crc kubenswrapper[4714]: I0129 16:37:27.356303 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qdbvx" Jan 29 16:37:27 crc kubenswrapper[4714]: I0129 16:37:27.358501 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qdbvx" Jan 29 16:37:27 crc kubenswrapper[4714]: I0129 16:37:27.408619 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qdbvx" Jan 29 16:37:27 crc kubenswrapper[4714]: I0129 16:37:27.844781 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:37:27 crc kubenswrapper[4714]: I0129 16:37:27.845164 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:37:27 crc kubenswrapper[4714]: I0129 16:37:27.869765 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qdbvx" Jan 29 16:37:27 crc kubenswrapper[4714]: I0129 16:37:27.918533 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qdbvx"] Jan 29 16:37:29 crc kubenswrapper[4714]: I0129 16:37:29.822553 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qdbvx" podUID="4e6b3552-9d7a-459f-a4c4-67da40c67c15" containerName="registry-server" containerID="cri-o://9632c9e5f7fc6621d6e14cceb0240050190e925aa5472c826d55f03bda4cac18" gracePeriod=2 Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.335790 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qdbvx" Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.391962 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7b4g\" (UniqueName: \"kubernetes.io/projected/4e6b3552-9d7a-459f-a4c4-67da40c67c15-kube-api-access-j7b4g\") pod \"4e6b3552-9d7a-459f-a4c4-67da40c67c15\" (UID: \"4e6b3552-9d7a-459f-a4c4-67da40c67c15\") " Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.392044 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e6b3552-9d7a-459f-a4c4-67da40c67c15-utilities\") pod \"4e6b3552-9d7a-459f-a4c4-67da40c67c15\" (UID: \"4e6b3552-9d7a-459f-a4c4-67da40c67c15\") " Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.392111 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e6b3552-9d7a-459f-a4c4-67da40c67c15-catalog-content\") pod \"4e6b3552-9d7a-459f-a4c4-67da40c67c15\" (UID: \"4e6b3552-9d7a-459f-a4c4-67da40c67c15\") " Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.393040 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e6b3552-9d7a-459f-a4c4-67da40c67c15-utilities" (OuterVolumeSpecName: "utilities") pod "4e6b3552-9d7a-459f-a4c4-67da40c67c15" (UID: "4e6b3552-9d7a-459f-a4c4-67da40c67c15"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.397168 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e6b3552-9d7a-459f-a4c4-67da40c67c15-kube-api-access-j7b4g" (OuterVolumeSpecName: "kube-api-access-j7b4g") pod "4e6b3552-9d7a-459f-a4c4-67da40c67c15" (UID: "4e6b3552-9d7a-459f-a4c4-67da40c67c15"). InnerVolumeSpecName "kube-api-access-j7b4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.445465 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e6b3552-9d7a-459f-a4c4-67da40c67c15-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e6b3552-9d7a-459f-a4c4-67da40c67c15" (UID: "4e6b3552-9d7a-459f-a4c4-67da40c67c15"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.493529 4714 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e6b3552-9d7a-459f-a4c4-67da40c67c15-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.493565 4714 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e6b3552-9d7a-459f-a4c4-67da40c67c15-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.493580 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7b4g\" (UniqueName: \"kubernetes.io/projected/4e6b3552-9d7a-459f-a4c4-67da40c67c15-kube-api-access-j7b4g\") on node \"crc\" DevicePath \"\"" Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.829895 4714 generic.go:334] "Generic (PLEG): container finished" podID="4e6b3552-9d7a-459f-a4c4-67da40c67c15" containerID="9632c9e5f7fc6621d6e14cceb0240050190e925aa5472c826d55f03bda4cac18" exitCode=0 Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.829952 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qdbvx" event={"ID":"4e6b3552-9d7a-459f-a4c4-67da40c67c15","Type":"ContainerDied","Data":"9632c9e5f7fc6621d6e14cceb0240050190e925aa5472c826d55f03bda4cac18"} Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.829983 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qdbvx" event={"ID":"4e6b3552-9d7a-459f-a4c4-67da40c67c15","Type":"ContainerDied","Data":"7cbae5276fb8d757c39c5f0cacfa9727ddbf2eb3306d3df09a254f9f755544f8"} Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.830000 4714 scope.go:117] "RemoveContainer" containerID="9632c9e5f7fc6621d6e14cceb0240050190e925aa5472c826d55f03bda4cac18" Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.830059 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qdbvx" Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.846306 4714 scope.go:117] "RemoveContainer" containerID="cd702a907a78fab939fb238809b451a3868bf29847022a3fd0a1279c986cdaca" Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.858616 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qdbvx"] Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.861959 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qdbvx"] Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.872639 4714 scope.go:117] "RemoveContainer" containerID="654e4e9ffb535db9fca5c14eda678189ef4ed63a10878793979a5fe56e16938d" Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.886710 4714 scope.go:117] "RemoveContainer" containerID="9632c9e5f7fc6621d6e14cceb0240050190e925aa5472c826d55f03bda4cac18" Jan 29 16:37:30 crc kubenswrapper[4714]: E0129 16:37:30.887198 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9632c9e5f7fc6621d6e14cceb0240050190e925aa5472c826d55f03bda4cac18\": container with ID starting with 9632c9e5f7fc6621d6e14cceb0240050190e925aa5472c826d55f03bda4cac18 not found: ID does not exist" containerID="9632c9e5f7fc6621d6e14cceb0240050190e925aa5472c826d55f03bda4cac18" Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.887237 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9632c9e5f7fc6621d6e14cceb0240050190e925aa5472c826d55f03bda4cac18"} err="failed to get container status \"9632c9e5f7fc6621d6e14cceb0240050190e925aa5472c826d55f03bda4cac18\": rpc error: code = NotFound desc = could not find container \"9632c9e5f7fc6621d6e14cceb0240050190e925aa5472c826d55f03bda4cac18\": container with ID starting with 9632c9e5f7fc6621d6e14cceb0240050190e925aa5472c826d55f03bda4cac18 not found: ID does not exist" Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.887266 4714 scope.go:117] "RemoveContainer" containerID="cd702a907a78fab939fb238809b451a3868bf29847022a3fd0a1279c986cdaca" Jan 29 16:37:30 crc kubenswrapper[4714]: E0129 16:37:30.887514 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd702a907a78fab939fb238809b451a3868bf29847022a3fd0a1279c986cdaca\": container with ID starting with cd702a907a78fab939fb238809b451a3868bf29847022a3fd0a1279c986cdaca not found: ID does not exist" containerID="cd702a907a78fab939fb238809b451a3868bf29847022a3fd0a1279c986cdaca" Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.887572 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd702a907a78fab939fb238809b451a3868bf29847022a3fd0a1279c986cdaca"} err="failed to get container status \"cd702a907a78fab939fb238809b451a3868bf29847022a3fd0a1279c986cdaca\": rpc error: code = NotFound desc = could not find container \"cd702a907a78fab939fb238809b451a3868bf29847022a3fd0a1279c986cdaca\": container with ID starting with cd702a907a78fab939fb238809b451a3868bf29847022a3fd0a1279c986cdaca not found: ID does not exist" Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.887615 4714 scope.go:117] "RemoveContainer" containerID="654e4e9ffb535db9fca5c14eda678189ef4ed63a10878793979a5fe56e16938d" Jan 29 16:37:30 crc kubenswrapper[4714]: E0129 16:37:30.887950 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"654e4e9ffb535db9fca5c14eda678189ef4ed63a10878793979a5fe56e16938d\": container with ID starting with 654e4e9ffb535db9fca5c14eda678189ef4ed63a10878793979a5fe56e16938d not found: ID does not exist" containerID="654e4e9ffb535db9fca5c14eda678189ef4ed63a10878793979a5fe56e16938d" Jan 29 16:37:30 crc kubenswrapper[4714]: I0129 16:37:30.887973 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"654e4e9ffb535db9fca5c14eda678189ef4ed63a10878793979a5fe56e16938d"} err="failed to get container status \"654e4e9ffb535db9fca5c14eda678189ef4ed63a10878793979a5fe56e16938d\": rpc error: code = NotFound desc = could not find container \"654e4e9ffb535db9fca5c14eda678189ef4ed63a10878793979a5fe56e16938d\": container with ID starting with 654e4e9ffb535db9fca5c14eda678189ef4ed63a10878793979a5fe56e16938d not found: ID does not exist" Jan 29 16:37:32 crc kubenswrapper[4714]: I0129 16:37:32.193166 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e6b3552-9d7a-459f-a4c4-67da40c67c15" path="/var/lib/kubelet/pods/4e6b3552-9d7a-459f-a4c4-67da40c67c15/volumes" Jan 29 16:37:33 crc kubenswrapper[4714]: I0129 16:37:33.953012 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-sq9mx_8062d225-aa57-48df-bf28-2254ecc4f635/control-plane-machine-set-operator/0.log" Jan 29 16:37:34 crc kubenswrapper[4714]: I0129 16:37:34.095352 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-z4h55_bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92/kube-rbac-proxy/0.log" Jan 29 16:37:34 crc kubenswrapper[4714]: I0129 16:37:34.122247 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-z4h55_bfb0bd22-cbd8-4ce8-a4f6-86a16dcdeb92/machine-api-operator/0.log" Jan 29 16:37:57 crc kubenswrapper[4714]: I0129 16:37:57.844835 4714 patch_prober.go:28] interesting pod/machine-config-daemon-ppngk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:37:57 crc kubenswrapper[4714]: I0129 16:37:57.845432 4714 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:37:57 crc kubenswrapper[4714]: I0129 16:37:57.845480 4714 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" Jan 29 16:37:57 crc kubenswrapper[4714]: I0129 16:37:57.846224 4714 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1"} pod="openshift-machine-config-operator/machine-config-daemon-ppngk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 16:37:57 crc kubenswrapper[4714]: I0129 16:37:57.846290 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerName="machine-config-daemon" containerID="cri-o://0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" gracePeriod=600 Jan 29 16:37:58 crc kubenswrapper[4714]: E0129 16:37:58.756546 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:37:59 crc kubenswrapper[4714]: I0129 16:37:59.001228 4714 generic.go:334] "Generic (PLEG): container finished" podID="c8c765f3-89eb-4077-8829-03e86eb0c90c" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" exitCode=0 Jan 29 16:37:59 crc kubenswrapper[4714]: I0129 16:37:59.001279 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" event={"ID":"c8c765f3-89eb-4077-8829-03e86eb0c90c","Type":"ContainerDied","Data":"0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1"} Jan 29 16:37:59 crc kubenswrapper[4714]: I0129 16:37:59.001330 4714 scope.go:117] "RemoveContainer" containerID="28ae6797628a288c954e7899195453697f32c3fca947d19910c2ccc63b246a5b" Jan 29 16:37:59 crc kubenswrapper[4714]: I0129 16:37:59.001957 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:37:59 crc kubenswrapper[4714]: E0129 16:37:59.002281 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:38:01 crc kubenswrapper[4714]: I0129 16:38:01.193668 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-m26zh_78b34628-144f-416a-b493-15ba445caa48/kube-rbac-proxy/0.log" Jan 29 16:38:01 crc kubenswrapper[4714]: I0129 16:38:01.202433 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-m26zh_78b34628-144f-416a-b493-15ba445caa48/controller/0.log" Jan 29 16:38:01 crc kubenswrapper[4714]: I0129 16:38:01.372732 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-frr-files/0.log" Jan 29 16:38:01 crc kubenswrapper[4714]: I0129 16:38:01.548182 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-reloader/0.log" Jan 29 16:38:01 crc kubenswrapper[4714]: I0129 16:38:01.557141 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-frr-files/0.log" Jan 29 16:38:01 crc kubenswrapper[4714]: I0129 16:38:01.624635 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-metrics/0.log" Jan 29 16:38:01 crc kubenswrapper[4714]: I0129 16:38:01.647768 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-reloader/0.log" Jan 29 16:38:01 crc kubenswrapper[4714]: I0129 16:38:01.738085 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-frr-files/0.log" Jan 29 16:38:01 crc kubenswrapper[4714]: I0129 16:38:01.764995 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-reloader/0.log" Jan 29 16:38:01 crc kubenswrapper[4714]: I0129 16:38:01.777148 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-metrics/0.log" Jan 29 16:38:01 crc kubenswrapper[4714]: I0129 16:38:01.821609 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-metrics/0.log" Jan 29 16:38:01 crc kubenswrapper[4714]: I0129 16:38:01.989901 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-reloader/0.log" Jan 29 16:38:02 crc kubenswrapper[4714]: I0129 16:38:02.000313 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-frr-files/0.log" Jan 29 16:38:02 crc kubenswrapper[4714]: I0129 16:38:02.042994 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/controller/0.log" Jan 29 16:38:02 crc kubenswrapper[4714]: I0129 16:38:02.064895 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/cp-metrics/0.log" Jan 29 16:38:02 crc kubenswrapper[4714]: I0129 16:38:02.183350 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/frr-metrics/0.log" Jan 29 16:38:02 crc kubenswrapper[4714]: I0129 16:38:02.214474 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/kube-rbac-proxy/0.log" Jan 29 16:38:02 crc kubenswrapper[4714]: I0129 16:38:02.300570 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/kube-rbac-proxy-frr/0.log" Jan 29 16:38:02 crc kubenswrapper[4714]: I0129 16:38:02.421580 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/reloader/0.log" Jan 29 16:38:02 crc kubenswrapper[4714]: I0129 16:38:02.508537 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-kk79r_9bbfcf92-8a27-4ba0-9017-7c36906791c8/frr-k8s-webhook-server/0.log" Jan 29 16:38:02 crc kubenswrapper[4714]: I0129 16:38:02.554095 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-59pmz_a97dd473-5873-4aa1-9166-f7a0c6581be1/frr/0.log" Jan 29 16:38:02 crc kubenswrapper[4714]: I0129 16:38:02.603703 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-586b87b897-zpr4q_432a4f98-877c-4f7a-b2b0-ce273a77450a/manager/0.log" Jan 29 16:38:02 crc kubenswrapper[4714]: I0129 16:38:02.714049 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7df7c8d444-xs67n_ffe179b8-a1c8-430b-94f5-920aacf0defe/webhook-server/0.log" Jan 29 16:38:02 crc kubenswrapper[4714]: I0129 16:38:02.832434 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7mmsh_813f735d-8336-49e9-b018-e6dbf74ddc99/kube-rbac-proxy/0.log" Jan 29 16:38:02 crc kubenswrapper[4714]: I0129 16:38:02.966695 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7mmsh_813f735d-8336-49e9-b018-e6dbf74ddc99/speaker/0.log" Jan 29 16:38:10 crc kubenswrapper[4714]: I0129 16:38:10.184312 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:38:10 crc kubenswrapper[4714]: E0129 16:38:10.184999 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:38:24 crc kubenswrapper[4714]: I0129 16:38:24.187053 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:38:24 crc kubenswrapper[4714]: E0129 16:38:24.187771 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:38:27 crc kubenswrapper[4714]: I0129 16:38:27.661892 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_c184c6f2-1af5-4f70-9251-6beb2baae06b/util/0.log" Jan 29 16:38:27 crc kubenswrapper[4714]: I0129 16:38:27.872702 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_c184c6f2-1af5-4f70-9251-6beb2baae06b/pull/0.log" Jan 29 16:38:27 crc kubenswrapper[4714]: I0129 16:38:27.878839 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_c184c6f2-1af5-4f70-9251-6beb2baae06b/pull/0.log" Jan 29 16:38:27 crc kubenswrapper[4714]: I0129 16:38:27.922982 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_c184c6f2-1af5-4f70-9251-6beb2baae06b/util/0.log" Jan 29 16:38:28 crc kubenswrapper[4714]: I0129 16:38:28.070011 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_c184c6f2-1af5-4f70-9251-6beb2baae06b/pull/0.log" Jan 29 16:38:28 crc kubenswrapper[4714]: I0129 16:38:28.086140 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_c184c6f2-1af5-4f70-9251-6beb2baae06b/extract/0.log" Jan 29 16:38:28 crc kubenswrapper[4714]: I0129 16:38:28.104407 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dclc9st_c184c6f2-1af5-4f70-9251-6beb2baae06b/util/0.log" Jan 29 16:38:28 crc kubenswrapper[4714]: I0129 16:38:28.251879 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slsxz_16cb244c-6c63-47e6-a312-ba33ab4d4899/extract-utilities/0.log" Jan 29 16:38:28 crc kubenswrapper[4714]: I0129 16:38:28.430417 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slsxz_16cb244c-6c63-47e6-a312-ba33ab4d4899/extract-content/0.log" Jan 29 16:38:28 crc kubenswrapper[4714]: I0129 16:38:28.434516 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slsxz_16cb244c-6c63-47e6-a312-ba33ab4d4899/extract-content/0.log" Jan 29 16:38:28 crc kubenswrapper[4714]: I0129 16:38:28.434827 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slsxz_16cb244c-6c63-47e6-a312-ba33ab4d4899/extract-utilities/0.log" Jan 29 16:38:28 crc kubenswrapper[4714]: I0129 16:38:28.631992 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slsxz_16cb244c-6c63-47e6-a312-ba33ab4d4899/extract-utilities/0.log" Jan 29 16:38:28 crc kubenswrapper[4714]: I0129 16:38:28.649009 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slsxz_16cb244c-6c63-47e6-a312-ba33ab4d4899/extract-content/0.log" Jan 29 16:38:28 crc kubenswrapper[4714]: I0129 16:38:28.817960 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ndx6p_ca655e22-8f97-4e9e-b115-734ae1af7d50/extract-utilities/0.log" Jan 29 16:38:28 crc kubenswrapper[4714]: I0129 16:38:28.977165 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ndx6p_ca655e22-8f97-4e9e-b115-734ae1af7d50/extract-utilities/0.log" Jan 29 16:38:28 crc kubenswrapper[4714]: I0129 16:38:28.985211 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ndx6p_ca655e22-8f97-4e9e-b115-734ae1af7d50/extract-content/0.log" Jan 29 16:38:29 crc kubenswrapper[4714]: I0129 16:38:29.043192 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slsxz_16cb244c-6c63-47e6-a312-ba33ab4d4899/registry-server/0.log" Jan 29 16:38:29 crc kubenswrapper[4714]: I0129 16:38:29.089312 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ndx6p_ca655e22-8f97-4e9e-b115-734ae1af7d50/extract-content/0.log" Jan 29 16:38:29 crc kubenswrapper[4714]: I0129 16:38:29.182056 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ndx6p_ca655e22-8f97-4e9e-b115-734ae1af7d50/extract-utilities/0.log" Jan 29 16:38:29 crc kubenswrapper[4714]: I0129 16:38:29.182085 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ndx6p_ca655e22-8f97-4e9e-b115-734ae1af7d50/extract-content/0.log" Jan 29 16:38:29 crc kubenswrapper[4714]: I0129 16:38:29.373777 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-7rvrl_2696757f-83ca-42df-9855-f76adeee02bb/marketplace-operator/0.log" Jan 29 16:38:29 crc kubenswrapper[4714]: I0129 16:38:29.483090 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6gkpz_04dba3a0-a89b-48c5-97ef-e5660d1ae7bb/extract-utilities/0.log" Jan 29 16:38:29 crc kubenswrapper[4714]: I0129 16:38:29.552731 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ndx6p_ca655e22-8f97-4e9e-b115-734ae1af7d50/registry-server/0.log" Jan 29 16:38:29 crc kubenswrapper[4714]: I0129 16:38:29.678451 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6gkpz_04dba3a0-a89b-48c5-97ef-e5660d1ae7bb/extract-utilities/0.log" Jan 29 16:38:29 crc kubenswrapper[4714]: I0129 16:38:29.686570 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6gkpz_04dba3a0-a89b-48c5-97ef-e5660d1ae7bb/extract-content/0.log" Jan 29 16:38:29 crc kubenswrapper[4714]: I0129 16:38:29.729897 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6gkpz_04dba3a0-a89b-48c5-97ef-e5660d1ae7bb/extract-content/0.log" Jan 29 16:38:29 crc kubenswrapper[4714]: I0129 16:38:29.852564 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6gkpz_04dba3a0-a89b-48c5-97ef-e5660d1ae7bb/extract-utilities/0.log" Jan 29 16:38:29 crc kubenswrapper[4714]: I0129 16:38:29.852564 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6gkpz_04dba3a0-a89b-48c5-97ef-e5660d1ae7bb/extract-content/0.log" Jan 29 16:38:29 crc kubenswrapper[4714]: I0129 16:38:29.959647 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6gkpz_04dba3a0-a89b-48c5-97ef-e5660d1ae7bb/registry-server/0.log" Jan 29 16:38:30 crc kubenswrapper[4714]: I0129 16:38:30.015760 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-knxc8_de6c9fbd-8657-4434-bff5-468276791466/extract-utilities/0.log" Jan 29 16:38:30 crc kubenswrapper[4714]: I0129 16:38:30.176002 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-knxc8_de6c9fbd-8657-4434-bff5-468276791466/extract-content/0.log" Jan 29 16:38:30 crc kubenswrapper[4714]: I0129 16:38:30.181216 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-knxc8_de6c9fbd-8657-4434-bff5-468276791466/extract-utilities/0.log" Jan 29 16:38:30 crc kubenswrapper[4714]: I0129 16:38:30.191205 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-knxc8_de6c9fbd-8657-4434-bff5-468276791466/extract-content/0.log" Jan 29 16:38:30 crc kubenswrapper[4714]: I0129 16:38:30.360054 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-knxc8_de6c9fbd-8657-4434-bff5-468276791466/extract-utilities/0.log" Jan 29 16:38:30 crc kubenswrapper[4714]: I0129 16:38:30.360360 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-knxc8_de6c9fbd-8657-4434-bff5-468276791466/extract-content/0.log" Jan 29 16:38:30 crc kubenswrapper[4714]: I0129 16:38:30.558424 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-knxc8_de6c9fbd-8657-4434-bff5-468276791466/registry-server/0.log" Jan 29 16:38:39 crc kubenswrapper[4714]: I0129 16:38:39.184127 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:38:39 crc kubenswrapper[4714]: E0129 16:38:39.184817 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:38:50 crc kubenswrapper[4714]: I0129 16:38:50.184239 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:38:50 crc kubenswrapper[4714]: E0129 16:38:50.185169 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:39:03 crc kubenswrapper[4714]: I0129 16:39:03.184868 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:39:03 crc kubenswrapper[4714]: E0129 16:39:03.185913 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:39:15 crc kubenswrapper[4714]: I0129 16:39:15.184754 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:39:15 crc kubenswrapper[4714]: E0129 16:39:15.185314 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:39:27 crc kubenswrapper[4714]: I0129 16:39:27.184239 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:39:27 crc kubenswrapper[4714]: E0129 16:39:27.185426 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:39:38 crc kubenswrapper[4714]: I0129 16:39:38.184144 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:39:38 crc kubenswrapper[4714]: E0129 16:39:38.184769 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:39:41 crc kubenswrapper[4714]: I0129 16:39:41.657243 4714 generic.go:334] "Generic (PLEG): container finished" podID="ff3dd7de-0ec3-4550-ba38-0d67405a2671" containerID="9fd93823c7fddefb604b23f3bb0a407aa212a848b3dc9ccb4d9664567595015f" exitCode=0 Jan 29 16:39:41 crc kubenswrapper[4714]: I0129 16:39:41.657328 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hqc4j/must-gather-684p8" event={"ID":"ff3dd7de-0ec3-4550-ba38-0d67405a2671","Type":"ContainerDied","Data":"9fd93823c7fddefb604b23f3bb0a407aa212a848b3dc9ccb4d9664567595015f"} Jan 29 16:39:41 crc kubenswrapper[4714]: I0129 16:39:41.658128 4714 scope.go:117] "RemoveContainer" containerID="9fd93823c7fddefb604b23f3bb0a407aa212a848b3dc9ccb4d9664567595015f" Jan 29 16:39:42 crc kubenswrapper[4714]: I0129 16:39:42.162224 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hqc4j_must-gather-684p8_ff3dd7de-0ec3-4550-ba38-0d67405a2671/gather/0.log" Jan 29 16:39:51 crc kubenswrapper[4714]: I0129 16:39:51.184467 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:39:51 crc kubenswrapper[4714]: E0129 16:39:51.186695 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:39:51 crc kubenswrapper[4714]: I0129 16:39:51.557998 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hqc4j/must-gather-684p8"] Jan 29 16:39:51 crc kubenswrapper[4714]: I0129 16:39:51.558514 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-hqc4j/must-gather-684p8" podUID="ff3dd7de-0ec3-4550-ba38-0d67405a2671" containerName="copy" containerID="cri-o://3660194f8ee015e6e72c5a402b3814eb8d75a532bbd31644dbe43ccd17971ac5" gracePeriod=2 Jan 29 16:39:51 crc kubenswrapper[4714]: I0129 16:39:51.563746 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hqc4j/must-gather-684p8"] Jan 29 16:39:52 crc kubenswrapper[4714]: I0129 16:39:52.459589 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hqc4j_must-gather-684p8_ff3dd7de-0ec3-4550-ba38-0d67405a2671/copy/0.log" Jan 29 16:39:52 crc kubenswrapper[4714]: I0129 16:39:52.460499 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hqc4j/must-gather-684p8" Jan 29 16:39:52 crc kubenswrapper[4714]: I0129 16:39:52.490359 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ff3dd7de-0ec3-4550-ba38-0d67405a2671-must-gather-output\") pod \"ff3dd7de-0ec3-4550-ba38-0d67405a2671\" (UID: \"ff3dd7de-0ec3-4550-ba38-0d67405a2671\") " Jan 29 16:39:52 crc kubenswrapper[4714]: I0129 16:39:52.490564 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jf26\" (UniqueName: \"kubernetes.io/projected/ff3dd7de-0ec3-4550-ba38-0d67405a2671-kube-api-access-8jf26\") pod \"ff3dd7de-0ec3-4550-ba38-0d67405a2671\" (UID: \"ff3dd7de-0ec3-4550-ba38-0d67405a2671\") " Jan 29 16:39:52 crc kubenswrapper[4714]: I0129 16:39:52.495348 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff3dd7de-0ec3-4550-ba38-0d67405a2671-kube-api-access-8jf26" (OuterVolumeSpecName: "kube-api-access-8jf26") pod "ff3dd7de-0ec3-4550-ba38-0d67405a2671" (UID: "ff3dd7de-0ec3-4550-ba38-0d67405a2671"). InnerVolumeSpecName "kube-api-access-8jf26". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:39:52 crc kubenswrapper[4714]: I0129 16:39:52.555027 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff3dd7de-0ec3-4550-ba38-0d67405a2671-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "ff3dd7de-0ec3-4550-ba38-0d67405a2671" (UID: "ff3dd7de-0ec3-4550-ba38-0d67405a2671"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:39:52 crc kubenswrapper[4714]: I0129 16:39:52.592349 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jf26\" (UniqueName: \"kubernetes.io/projected/ff3dd7de-0ec3-4550-ba38-0d67405a2671-kube-api-access-8jf26\") on node \"crc\" DevicePath \"\"" Jan 29 16:39:52 crc kubenswrapper[4714]: I0129 16:39:52.592402 4714 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ff3dd7de-0ec3-4550-ba38-0d67405a2671-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 29 16:39:52 crc kubenswrapper[4714]: I0129 16:39:52.733091 4714 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hqc4j_must-gather-684p8_ff3dd7de-0ec3-4550-ba38-0d67405a2671/copy/0.log" Jan 29 16:39:52 crc kubenswrapper[4714]: I0129 16:39:52.733796 4714 generic.go:334] "Generic (PLEG): container finished" podID="ff3dd7de-0ec3-4550-ba38-0d67405a2671" containerID="3660194f8ee015e6e72c5a402b3814eb8d75a532bbd31644dbe43ccd17971ac5" exitCode=143 Jan 29 16:39:52 crc kubenswrapper[4714]: I0129 16:39:52.733842 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hqc4j/must-gather-684p8" Jan 29 16:39:52 crc kubenswrapper[4714]: I0129 16:39:52.733857 4714 scope.go:117] "RemoveContainer" containerID="3660194f8ee015e6e72c5a402b3814eb8d75a532bbd31644dbe43ccd17971ac5" Jan 29 16:39:52 crc kubenswrapper[4714]: I0129 16:39:52.749527 4714 scope.go:117] "RemoveContainer" containerID="9fd93823c7fddefb604b23f3bb0a407aa212a848b3dc9ccb4d9664567595015f" Jan 29 16:39:52 crc kubenswrapper[4714]: I0129 16:39:52.804095 4714 scope.go:117] "RemoveContainer" containerID="3660194f8ee015e6e72c5a402b3814eb8d75a532bbd31644dbe43ccd17971ac5" Jan 29 16:39:52 crc kubenswrapper[4714]: E0129 16:39:52.804490 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3660194f8ee015e6e72c5a402b3814eb8d75a532bbd31644dbe43ccd17971ac5\": container with ID starting with 3660194f8ee015e6e72c5a402b3814eb8d75a532bbd31644dbe43ccd17971ac5 not found: ID does not exist" containerID="3660194f8ee015e6e72c5a402b3814eb8d75a532bbd31644dbe43ccd17971ac5" Jan 29 16:39:52 crc kubenswrapper[4714]: I0129 16:39:52.804531 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3660194f8ee015e6e72c5a402b3814eb8d75a532bbd31644dbe43ccd17971ac5"} err="failed to get container status \"3660194f8ee015e6e72c5a402b3814eb8d75a532bbd31644dbe43ccd17971ac5\": rpc error: code = NotFound desc = could not find container \"3660194f8ee015e6e72c5a402b3814eb8d75a532bbd31644dbe43ccd17971ac5\": container with ID starting with 3660194f8ee015e6e72c5a402b3814eb8d75a532bbd31644dbe43ccd17971ac5 not found: ID does not exist" Jan 29 16:39:52 crc kubenswrapper[4714]: I0129 16:39:52.804557 4714 scope.go:117] "RemoveContainer" containerID="9fd93823c7fddefb604b23f3bb0a407aa212a848b3dc9ccb4d9664567595015f" Jan 29 16:39:52 crc kubenswrapper[4714]: E0129 16:39:52.805006 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fd93823c7fddefb604b23f3bb0a407aa212a848b3dc9ccb4d9664567595015f\": container with ID starting with 9fd93823c7fddefb604b23f3bb0a407aa212a848b3dc9ccb4d9664567595015f not found: ID does not exist" containerID="9fd93823c7fddefb604b23f3bb0a407aa212a848b3dc9ccb4d9664567595015f" Jan 29 16:39:52 crc kubenswrapper[4714]: I0129 16:39:52.805041 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fd93823c7fddefb604b23f3bb0a407aa212a848b3dc9ccb4d9664567595015f"} err="failed to get container status \"9fd93823c7fddefb604b23f3bb0a407aa212a848b3dc9ccb4d9664567595015f\": rpc error: code = NotFound desc = could not find container \"9fd93823c7fddefb604b23f3bb0a407aa212a848b3dc9ccb4d9664567595015f\": container with ID starting with 9fd93823c7fddefb604b23f3bb0a407aa212a848b3dc9ccb4d9664567595015f not found: ID does not exist" Jan 29 16:39:54 crc kubenswrapper[4714]: I0129 16:39:54.201342 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff3dd7de-0ec3-4550-ba38-0d67405a2671" path="/var/lib/kubelet/pods/ff3dd7de-0ec3-4550-ba38-0d67405a2671/volumes" Jan 29 16:40:06 crc kubenswrapper[4714]: I0129 16:40:06.184714 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:40:06 crc kubenswrapper[4714]: E0129 16:40:06.185900 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:40:18 crc kubenswrapper[4714]: I0129 16:40:18.184499 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:40:18 crc kubenswrapper[4714]: E0129 16:40:18.185391 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:40:31 crc kubenswrapper[4714]: I0129 16:40:31.185129 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:40:31 crc kubenswrapper[4714]: E0129 16:40:31.186449 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.625224 4714 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x6vt6"] Jan 29 16:40:41 crc kubenswrapper[4714]: E0129 16:40:41.626496 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff3dd7de-0ec3-4550-ba38-0d67405a2671" containerName="gather" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.626520 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff3dd7de-0ec3-4550-ba38-0d67405a2671" containerName="gather" Jan 29 16:40:41 crc kubenswrapper[4714]: E0129 16:40:41.626540 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f29ce50e-bbd2-4fe5-a032-5163f3f80ca0" containerName="extract-utilities" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.626553 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="f29ce50e-bbd2-4fe5-a032-5163f3f80ca0" containerName="extract-utilities" Jan 29 16:40:41 crc kubenswrapper[4714]: E0129 16:40:41.626577 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e6b3552-9d7a-459f-a4c4-67da40c67c15" containerName="registry-server" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.626592 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e6b3552-9d7a-459f-a4c4-67da40c67c15" containerName="registry-server" Jan 29 16:40:41 crc kubenswrapper[4714]: E0129 16:40:41.626605 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff3dd7de-0ec3-4550-ba38-0d67405a2671" containerName="copy" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.626618 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff3dd7de-0ec3-4550-ba38-0d67405a2671" containerName="copy" Jan 29 16:40:41 crc kubenswrapper[4714]: E0129 16:40:41.626642 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f29ce50e-bbd2-4fe5-a032-5163f3f80ca0" containerName="extract-content" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.626655 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="f29ce50e-bbd2-4fe5-a032-5163f3f80ca0" containerName="extract-content" Jan 29 16:40:41 crc kubenswrapper[4714]: E0129 16:40:41.626682 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e6b3552-9d7a-459f-a4c4-67da40c67c15" containerName="extract-utilities" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.626695 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e6b3552-9d7a-459f-a4c4-67da40c67c15" containerName="extract-utilities" Jan 29 16:40:41 crc kubenswrapper[4714]: E0129 16:40:41.626724 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e6b3552-9d7a-459f-a4c4-67da40c67c15" containerName="extract-content" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.626736 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e6b3552-9d7a-459f-a4c4-67da40c67c15" containerName="extract-content" Jan 29 16:40:41 crc kubenswrapper[4714]: E0129 16:40:41.626755 4714 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f29ce50e-bbd2-4fe5-a032-5163f3f80ca0" containerName="registry-server" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.626769 4714 state_mem.go:107] "Deleted CPUSet assignment" podUID="f29ce50e-bbd2-4fe5-a032-5163f3f80ca0" containerName="registry-server" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.627012 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff3dd7de-0ec3-4550-ba38-0d67405a2671" containerName="gather" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.627040 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="f29ce50e-bbd2-4fe5-a032-5163f3f80ca0" containerName="registry-server" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.627058 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e6b3552-9d7a-459f-a4c4-67da40c67c15" containerName="registry-server" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.627071 4714 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff3dd7de-0ec3-4550-ba38-0d67405a2671" containerName="copy" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.628527 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x6vt6" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.643169 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x6vt6"] Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.702213 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bb90f87-76b4-4b43-9242-5640128dbed9-utilities\") pod \"redhat-operators-x6vt6\" (UID: \"4bb90f87-76b4-4b43-9242-5640128dbed9\") " pod="openshift-marketplace/redhat-operators-x6vt6" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.702291 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bb90f87-76b4-4b43-9242-5640128dbed9-catalog-content\") pod \"redhat-operators-x6vt6\" (UID: \"4bb90f87-76b4-4b43-9242-5640128dbed9\") " pod="openshift-marketplace/redhat-operators-x6vt6" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.702338 4714 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f5nw\" (UniqueName: \"kubernetes.io/projected/4bb90f87-76b4-4b43-9242-5640128dbed9-kube-api-access-9f5nw\") pod \"redhat-operators-x6vt6\" (UID: \"4bb90f87-76b4-4b43-9242-5640128dbed9\") " pod="openshift-marketplace/redhat-operators-x6vt6" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.803247 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bb90f87-76b4-4b43-9242-5640128dbed9-utilities\") pod \"redhat-operators-x6vt6\" (UID: \"4bb90f87-76b4-4b43-9242-5640128dbed9\") " pod="openshift-marketplace/redhat-operators-x6vt6" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.803318 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bb90f87-76b4-4b43-9242-5640128dbed9-catalog-content\") pod \"redhat-operators-x6vt6\" (UID: \"4bb90f87-76b4-4b43-9242-5640128dbed9\") " pod="openshift-marketplace/redhat-operators-x6vt6" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.803366 4714 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f5nw\" (UniqueName: \"kubernetes.io/projected/4bb90f87-76b4-4b43-9242-5640128dbed9-kube-api-access-9f5nw\") pod \"redhat-operators-x6vt6\" (UID: \"4bb90f87-76b4-4b43-9242-5640128dbed9\") " pod="openshift-marketplace/redhat-operators-x6vt6" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.804126 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bb90f87-76b4-4b43-9242-5640128dbed9-utilities\") pod \"redhat-operators-x6vt6\" (UID: \"4bb90f87-76b4-4b43-9242-5640128dbed9\") " pod="openshift-marketplace/redhat-operators-x6vt6" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.804433 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bb90f87-76b4-4b43-9242-5640128dbed9-catalog-content\") pod \"redhat-operators-x6vt6\" (UID: \"4bb90f87-76b4-4b43-9242-5640128dbed9\") " pod="openshift-marketplace/redhat-operators-x6vt6" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.826883 4714 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f5nw\" (UniqueName: \"kubernetes.io/projected/4bb90f87-76b4-4b43-9242-5640128dbed9-kube-api-access-9f5nw\") pod \"redhat-operators-x6vt6\" (UID: \"4bb90f87-76b4-4b43-9242-5640128dbed9\") " pod="openshift-marketplace/redhat-operators-x6vt6" Jan 29 16:40:41 crc kubenswrapper[4714]: I0129 16:40:41.966846 4714 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x6vt6" Jan 29 16:40:42 crc kubenswrapper[4714]: I0129 16:40:42.173744 4714 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x6vt6"] Jan 29 16:40:43 crc kubenswrapper[4714]: I0129 16:40:43.068321 4714 generic.go:334] "Generic (PLEG): container finished" podID="4bb90f87-76b4-4b43-9242-5640128dbed9" containerID="e74a0b3f97545922aa02f04f1f78c55567e52209ac7863ae6e2cc02588e55ab7" exitCode=0 Jan 29 16:40:43 crc kubenswrapper[4714]: I0129 16:40:43.068366 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x6vt6" event={"ID":"4bb90f87-76b4-4b43-9242-5640128dbed9","Type":"ContainerDied","Data":"e74a0b3f97545922aa02f04f1f78c55567e52209ac7863ae6e2cc02588e55ab7"} Jan 29 16:40:43 crc kubenswrapper[4714]: I0129 16:40:43.068394 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x6vt6" event={"ID":"4bb90f87-76b4-4b43-9242-5640128dbed9","Type":"ContainerStarted","Data":"491231bcd133cd91582b60d1cec8a74ce694970727fe9817d84874a1355e3195"} Jan 29 16:40:45 crc kubenswrapper[4714]: I0129 16:40:45.083228 4714 generic.go:334] "Generic (PLEG): container finished" podID="4bb90f87-76b4-4b43-9242-5640128dbed9" containerID="e0be61f9d330e704f6ad4da95588bd54481951c4968e8ce1ae9312e631fe96f2" exitCode=0 Jan 29 16:40:45 crc kubenswrapper[4714]: I0129 16:40:45.083331 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x6vt6" event={"ID":"4bb90f87-76b4-4b43-9242-5640128dbed9","Type":"ContainerDied","Data":"e0be61f9d330e704f6ad4da95588bd54481951c4968e8ce1ae9312e631fe96f2"} Jan 29 16:40:46 crc kubenswrapper[4714]: I0129 16:40:46.092334 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x6vt6" event={"ID":"4bb90f87-76b4-4b43-9242-5640128dbed9","Type":"ContainerStarted","Data":"2cd4a929c1ff0695bb4467a8ea29639e901901262b6e43af866569c0d36a51c6"} Jan 29 16:40:46 crc kubenswrapper[4714]: I0129 16:40:46.117140 4714 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x6vt6" podStartSLOduration=2.447135732 podStartE2EDuration="5.117120243s" podCreationTimestamp="2026-01-29 16:40:41 +0000 UTC" firstStartedPulling="2026-01-29 16:40:43.070835758 +0000 UTC m=+1849.591336888" lastFinishedPulling="2026-01-29 16:40:45.740820249 +0000 UTC m=+1852.261321399" observedRunningTime="2026-01-29 16:40:46.1138142 +0000 UTC m=+1852.634315330" watchObservedRunningTime="2026-01-29 16:40:46.117120243 +0000 UTC m=+1852.637621363" Jan 29 16:40:46 crc kubenswrapper[4714]: I0129 16:40:46.185270 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:40:46 crc kubenswrapper[4714]: E0129 16:40:46.185523 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:40:51 crc kubenswrapper[4714]: I0129 16:40:51.967562 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x6vt6" Jan 29 16:40:51 crc kubenswrapper[4714]: I0129 16:40:51.968847 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x6vt6" Jan 29 16:40:53 crc kubenswrapper[4714]: I0129 16:40:53.046617 4714 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x6vt6" podUID="4bb90f87-76b4-4b43-9242-5640128dbed9" containerName="registry-server" probeResult="failure" output=< Jan 29 16:40:53 crc kubenswrapper[4714]: timeout: failed to connect service ":50051" within 1s Jan 29 16:40:53 crc kubenswrapper[4714]: > Jan 29 16:40:58 crc kubenswrapper[4714]: I0129 16:40:58.184029 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:40:58 crc kubenswrapper[4714]: E0129 16:40:58.184509 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:41:02 crc kubenswrapper[4714]: I0129 16:41:02.001778 4714 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x6vt6" Jan 29 16:41:02 crc kubenswrapper[4714]: I0129 16:41:02.039966 4714 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x6vt6" Jan 29 16:41:02 crc kubenswrapper[4714]: I0129 16:41:02.530543 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x6vt6"] Jan 29 16:41:03 crc kubenswrapper[4714]: I0129 16:41:03.218983 4714 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x6vt6" podUID="4bb90f87-76b4-4b43-9242-5640128dbed9" containerName="registry-server" containerID="cri-o://2cd4a929c1ff0695bb4467a8ea29639e901901262b6e43af866569c0d36a51c6" gracePeriod=2 Jan 29 16:41:03 crc kubenswrapper[4714]: I0129 16:41:03.652042 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x6vt6" Jan 29 16:41:03 crc kubenswrapper[4714]: I0129 16:41:03.821828 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bb90f87-76b4-4b43-9242-5640128dbed9-utilities\") pod \"4bb90f87-76b4-4b43-9242-5640128dbed9\" (UID: \"4bb90f87-76b4-4b43-9242-5640128dbed9\") " Jan 29 16:41:03 crc kubenswrapper[4714]: I0129 16:41:03.821872 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9f5nw\" (UniqueName: \"kubernetes.io/projected/4bb90f87-76b4-4b43-9242-5640128dbed9-kube-api-access-9f5nw\") pod \"4bb90f87-76b4-4b43-9242-5640128dbed9\" (UID: \"4bb90f87-76b4-4b43-9242-5640128dbed9\") " Jan 29 16:41:03 crc kubenswrapper[4714]: I0129 16:41:03.821893 4714 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bb90f87-76b4-4b43-9242-5640128dbed9-catalog-content\") pod \"4bb90f87-76b4-4b43-9242-5640128dbed9\" (UID: \"4bb90f87-76b4-4b43-9242-5640128dbed9\") " Jan 29 16:41:03 crc kubenswrapper[4714]: I0129 16:41:03.823125 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bb90f87-76b4-4b43-9242-5640128dbed9-utilities" (OuterVolumeSpecName: "utilities") pod "4bb90f87-76b4-4b43-9242-5640128dbed9" (UID: "4bb90f87-76b4-4b43-9242-5640128dbed9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:41:03 crc kubenswrapper[4714]: I0129 16:41:03.830986 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb90f87-76b4-4b43-9242-5640128dbed9-kube-api-access-9f5nw" (OuterVolumeSpecName: "kube-api-access-9f5nw") pod "4bb90f87-76b4-4b43-9242-5640128dbed9" (UID: "4bb90f87-76b4-4b43-9242-5640128dbed9"). InnerVolumeSpecName "kube-api-access-9f5nw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:41:03 crc kubenswrapper[4714]: I0129 16:41:03.924509 4714 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bb90f87-76b4-4b43-9242-5640128dbed9-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:41:03 crc kubenswrapper[4714]: I0129 16:41:03.924573 4714 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9f5nw\" (UniqueName: \"kubernetes.io/projected/4bb90f87-76b4-4b43-9242-5640128dbed9-kube-api-access-9f5nw\") on node \"crc\" DevicePath \"\"" Jan 29 16:41:03 crc kubenswrapper[4714]: I0129 16:41:03.998194 4714 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bb90f87-76b4-4b43-9242-5640128dbed9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4bb90f87-76b4-4b43-9242-5640128dbed9" (UID: "4bb90f87-76b4-4b43-9242-5640128dbed9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:41:04 crc kubenswrapper[4714]: I0129 16:41:04.025241 4714 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bb90f87-76b4-4b43-9242-5640128dbed9-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:41:04 crc kubenswrapper[4714]: I0129 16:41:04.231648 4714 generic.go:334] "Generic (PLEG): container finished" podID="4bb90f87-76b4-4b43-9242-5640128dbed9" containerID="2cd4a929c1ff0695bb4467a8ea29639e901901262b6e43af866569c0d36a51c6" exitCode=0 Jan 29 16:41:04 crc kubenswrapper[4714]: I0129 16:41:04.231699 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x6vt6" event={"ID":"4bb90f87-76b4-4b43-9242-5640128dbed9","Type":"ContainerDied","Data":"2cd4a929c1ff0695bb4467a8ea29639e901901262b6e43af866569c0d36a51c6"} Jan 29 16:41:04 crc kubenswrapper[4714]: I0129 16:41:04.231732 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x6vt6" event={"ID":"4bb90f87-76b4-4b43-9242-5640128dbed9","Type":"ContainerDied","Data":"491231bcd133cd91582b60d1cec8a74ce694970727fe9817d84874a1355e3195"} Jan 29 16:41:04 crc kubenswrapper[4714]: I0129 16:41:04.231756 4714 scope.go:117] "RemoveContainer" containerID="2cd4a929c1ff0695bb4467a8ea29639e901901262b6e43af866569c0d36a51c6" Jan 29 16:41:04 crc kubenswrapper[4714]: I0129 16:41:04.232128 4714 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x6vt6" Jan 29 16:41:04 crc kubenswrapper[4714]: I0129 16:41:04.255539 4714 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x6vt6"] Jan 29 16:41:04 crc kubenswrapper[4714]: I0129 16:41:04.259555 4714 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x6vt6"] Jan 29 16:41:04 crc kubenswrapper[4714]: I0129 16:41:04.272055 4714 scope.go:117] "RemoveContainer" containerID="e0be61f9d330e704f6ad4da95588bd54481951c4968e8ce1ae9312e631fe96f2" Jan 29 16:41:04 crc kubenswrapper[4714]: I0129 16:41:04.292478 4714 scope.go:117] "RemoveContainer" containerID="e74a0b3f97545922aa02f04f1f78c55567e52209ac7863ae6e2cc02588e55ab7" Jan 29 16:41:04 crc kubenswrapper[4714]: I0129 16:41:04.315267 4714 scope.go:117] "RemoveContainer" containerID="2cd4a929c1ff0695bb4467a8ea29639e901901262b6e43af866569c0d36a51c6" Jan 29 16:41:04 crc kubenswrapper[4714]: E0129 16:41:04.315826 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cd4a929c1ff0695bb4467a8ea29639e901901262b6e43af866569c0d36a51c6\": container with ID starting with 2cd4a929c1ff0695bb4467a8ea29639e901901262b6e43af866569c0d36a51c6 not found: ID does not exist" containerID="2cd4a929c1ff0695bb4467a8ea29639e901901262b6e43af866569c0d36a51c6" Jan 29 16:41:04 crc kubenswrapper[4714]: I0129 16:41:04.315876 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cd4a929c1ff0695bb4467a8ea29639e901901262b6e43af866569c0d36a51c6"} err="failed to get container status \"2cd4a929c1ff0695bb4467a8ea29639e901901262b6e43af866569c0d36a51c6\": rpc error: code = NotFound desc = could not find container \"2cd4a929c1ff0695bb4467a8ea29639e901901262b6e43af866569c0d36a51c6\": container with ID starting with 2cd4a929c1ff0695bb4467a8ea29639e901901262b6e43af866569c0d36a51c6 not found: ID does not exist" Jan 29 16:41:04 crc kubenswrapper[4714]: I0129 16:41:04.315919 4714 scope.go:117] "RemoveContainer" containerID="e0be61f9d330e704f6ad4da95588bd54481951c4968e8ce1ae9312e631fe96f2" Jan 29 16:41:04 crc kubenswrapper[4714]: E0129 16:41:04.316257 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0be61f9d330e704f6ad4da95588bd54481951c4968e8ce1ae9312e631fe96f2\": container with ID starting with e0be61f9d330e704f6ad4da95588bd54481951c4968e8ce1ae9312e631fe96f2 not found: ID does not exist" containerID="e0be61f9d330e704f6ad4da95588bd54481951c4968e8ce1ae9312e631fe96f2" Jan 29 16:41:04 crc kubenswrapper[4714]: I0129 16:41:04.316281 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0be61f9d330e704f6ad4da95588bd54481951c4968e8ce1ae9312e631fe96f2"} err="failed to get container status \"e0be61f9d330e704f6ad4da95588bd54481951c4968e8ce1ae9312e631fe96f2\": rpc error: code = NotFound desc = could not find container \"e0be61f9d330e704f6ad4da95588bd54481951c4968e8ce1ae9312e631fe96f2\": container with ID starting with e0be61f9d330e704f6ad4da95588bd54481951c4968e8ce1ae9312e631fe96f2 not found: ID does not exist" Jan 29 16:41:04 crc kubenswrapper[4714]: I0129 16:41:04.316294 4714 scope.go:117] "RemoveContainer" containerID="e74a0b3f97545922aa02f04f1f78c55567e52209ac7863ae6e2cc02588e55ab7" Jan 29 16:41:04 crc kubenswrapper[4714]: E0129 16:41:04.316521 4714 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e74a0b3f97545922aa02f04f1f78c55567e52209ac7863ae6e2cc02588e55ab7\": container with ID starting with e74a0b3f97545922aa02f04f1f78c55567e52209ac7863ae6e2cc02588e55ab7 not found: ID does not exist" containerID="e74a0b3f97545922aa02f04f1f78c55567e52209ac7863ae6e2cc02588e55ab7" Jan 29 16:41:04 crc kubenswrapper[4714]: I0129 16:41:04.316554 4714 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e74a0b3f97545922aa02f04f1f78c55567e52209ac7863ae6e2cc02588e55ab7"} err="failed to get container status \"e74a0b3f97545922aa02f04f1f78c55567e52209ac7863ae6e2cc02588e55ab7\": rpc error: code = NotFound desc = could not find container \"e74a0b3f97545922aa02f04f1f78c55567e52209ac7863ae6e2cc02588e55ab7\": container with ID starting with e74a0b3f97545922aa02f04f1f78c55567e52209ac7863ae6e2cc02588e55ab7 not found: ID does not exist" Jan 29 16:41:06 crc kubenswrapper[4714]: I0129 16:41:06.198474 4714 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb90f87-76b4-4b43-9242-5640128dbed9" path="/var/lib/kubelet/pods/4bb90f87-76b4-4b43-9242-5640128dbed9/volumes" Jan 29 16:41:10 crc kubenswrapper[4714]: I0129 16:41:10.184122 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:41:10 crc kubenswrapper[4714]: E0129 16:41:10.184953 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:41:24 crc kubenswrapper[4714]: I0129 16:41:24.190685 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:41:24 crc kubenswrapper[4714]: E0129 16:41:24.191829 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:41:39 crc kubenswrapper[4714]: I0129 16:41:39.183779 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:41:39 crc kubenswrapper[4714]: E0129 16:41:39.185021 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:41:51 crc kubenswrapper[4714]: I0129 16:41:51.184263 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:41:51 crc kubenswrapper[4714]: E0129 16:41:51.185542 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:42:02 crc kubenswrapper[4714]: I0129 16:42:02.184638 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:42:02 crc kubenswrapper[4714]: E0129 16:42:02.185641 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:42:17 crc kubenswrapper[4714]: I0129 16:42:17.184478 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:42:17 crc kubenswrapper[4714]: E0129 16:42:17.185426 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:42:30 crc kubenswrapper[4714]: I0129 16:42:30.183913 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:42:30 crc kubenswrapper[4714]: E0129 16:42:30.184885 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:42:42 crc kubenswrapper[4714]: I0129 16:42:42.184451 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:42:42 crc kubenswrapper[4714]: E0129 16:42:42.185453 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:42:54 crc kubenswrapper[4714]: I0129 16:42:54.188964 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:42:54 crc kubenswrapper[4714]: E0129 16:42:54.189713 4714 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ppngk_openshift-machine-config-operator(c8c765f3-89eb-4077-8829-03e86eb0c90c)\"" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" podUID="c8c765f3-89eb-4077-8829-03e86eb0c90c" Jan 29 16:43:06 crc kubenswrapper[4714]: I0129 16:43:06.193555 4714 scope.go:117] "RemoveContainer" containerID="0f064189d4746edc06d2b74e60a2cbff7511efd665171516db63a8272ebb29e1" Jan 29 16:43:07 crc kubenswrapper[4714]: I0129 16:43:07.114307 4714 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ppngk" event={"ID":"c8c765f3-89eb-4077-8829-03e86eb0c90c","Type":"ContainerStarted","Data":"2e6f448b0ca4f02ef396e0fff95441e590d42fc0999bf61b3886954adf50616e"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515136707113024451 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015136707113017366 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015136702615016513 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015136702615015463 5ustar corecore